[ 538.569704] env[62952]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62952) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 538.570084] env[62952]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62952) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 538.570185] env[62952]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62952) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 538.570501] env[62952]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 538.662957] env[62952]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62952) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 538.672667] env[62952]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62952) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 539.273790] env[62952]: INFO nova.virt.driver [None req-68bf184b-7df3-41a7-a7e5-19304cf0eeb4 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 539.344481] env[62952]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.344662] env[62952]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.344721] env[62952]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62952) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 542.471132] env[62952]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-d04d9acc-4723-4d30-b710-d73e523da9cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.486896] env[62952]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62952) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 542.487082] env[62952]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-6ec24ef5-abef-47e6-ae4c-422165e0e99b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.518447] env[62952]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 58b60. [ 542.518626] env[62952]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.174s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.519195] env[62952]: INFO nova.virt.vmwareapi.driver [None req-68bf184b-7df3-41a7-a7e5-19304cf0eeb4 None None] VMware vCenter version: 7.0.3 [ 542.522626] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0f9b9e-e2d1-4329-b21e-91f80fff4d9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.539673] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81aeeb7b-270e-406b-b5ff-9afb990460db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.545512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcef711e-0560-433c-b483-ff8843eeffc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.551950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016a34de-fed0-4532-97d2-ef478d39d498 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.564612] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a4edd6-e9d2-4e16-b31f-b1b30bf85bd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.570375] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b6e513-8c1f-483a-867e-48bbd4cb738c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.599757] env[62952]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-bb6d2bed-59f6-464e-92ff-2fa616866212 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.605008] env[62952]: DEBUG nova.virt.vmwareapi.driver [None req-68bf184b-7df3-41a7-a7e5-19304cf0eeb4 None None] Extension org.openstack.compute already exists. {{(pid=62952) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 542.607666] env[62952]: INFO nova.compute.provider_config [None req-68bf184b-7df3-41a7-a7e5-19304cf0eeb4 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 543.115022] env[62952]: DEBUG nova.context [None req-68bf184b-7df3-41a7-a7e5-19304cf0eeb4 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),c3910535-3dcf-4557-8dc9-54a14b20e0a3(cell1) {{(pid=62952) load_cells /opt/stack/nova/nova/context.py:464}} [ 543.115022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.115022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.115387] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.116023] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Acquiring lock "c3910535-3dcf-4557-8dc9-54a14b20e0a3" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.116371] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Lock "c3910535-3dcf-4557-8dc9-54a14b20e0a3" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.117542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Lock "c3910535-3dcf-4557-8dc9-54a14b20e0a3" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.137706] env[62952]: INFO dbcounter [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Registered counter for database nova_cell0 [ 543.146076] env[62952]: INFO dbcounter [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Registered counter for database nova_cell1 [ 543.149670] env[62952]: DEBUG oslo_db.sqlalchemy.engines [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62952) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 543.150150] env[62952]: DEBUG oslo_db.sqlalchemy.engines [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62952) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 543.155144] env[62952]: ERROR nova.db.main.api [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.155144] env[62952]: result = function(*args, **kwargs) [ 543.155144] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.155144] env[62952]: return func(*args, **kwargs) [ 543.155144] env[62952]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 543.155144] env[62952]: result = fn(*args, **kwargs) [ 543.155144] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 543.155144] env[62952]: return f(*args, **kwargs) [ 543.155144] env[62952]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 543.155144] env[62952]: return db.service_get_minimum_version(context, binaries) [ 543.155144] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 543.155144] env[62952]: _check_db_access() [ 543.155144] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 543.155144] env[62952]: stacktrace = ''.join(traceback.format_stack()) [ 543.155144] env[62952]: [ 543.157024] env[62952]: ERROR nova.db.main.api [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.157024] env[62952]: result = function(*args, **kwargs) [ 543.157024] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.157024] env[62952]: return func(*args, **kwargs) [ 543.157024] env[62952]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 543.157024] env[62952]: result = fn(*args, **kwargs) [ 543.157024] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 543.157024] env[62952]: return f(*args, **kwargs) [ 543.157024] env[62952]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 543.157024] env[62952]: return db.service_get_minimum_version(context, binaries) [ 543.157024] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 543.157024] env[62952]: _check_db_access() [ 543.157024] env[62952]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 543.157024] env[62952]: stacktrace = ''.join(traceback.format_stack()) [ 543.157024] env[62952]: [ 543.161194] env[62952]: WARNING nova.objects.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Failed to get minimum service version for cell c3910535-3dcf-4557-8dc9-54a14b20e0a3 [ 543.161194] env[62952]: WARNING nova.objects.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 543.161194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Acquiring lock "singleton_lock" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.161194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Acquired lock "singleton_lock" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.161194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Releasing lock "singleton_lock" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.161194] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Full set of CONF: {{(pid=62952) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 543.161194] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ******************************************************************************** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 543.161194] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Configuration options gathered from: {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 543.161395] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 543.161395] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 543.161395] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ================================================================================ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 543.161395] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] allow_resize_to_same_host = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161395] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] arq_binding_timeout = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161395] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] backdoor_port = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161556] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] backdoor_socket = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161556] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] block_device_allocate_retries = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161556] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] block_device_allocate_retries_interval = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161556] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cert = self.pem {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161556] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.161893] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute_monitors = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.162206] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] config_dir = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.162590] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] config_drive_format = iso9660 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] config_source = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] console_host = devstack {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] control_exchange = nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cpu_allocation_ratio = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] daemon = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.165957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] debug = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166212] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] default_access_ip_network_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166212] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] default_availability_zone = nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166212] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] default_ephemeral_format = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166212] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] default_green_pool_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166212] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] default_schedule_zone = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] disk_allocation_ratio = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] enable_new_services = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] enabled_apis = ['osapi_compute'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] enabled_ssl_apis = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] flat_injected = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] force_config_drive = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.166880] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] force_raw_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.167212] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] graceful_shutdown_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.167538] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] heal_instance_info_cache_interval = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.167896] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] host = cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.168220] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_build_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_delete_interval = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_format = [instance: %(uuid)s] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_name_template = instance-%08x {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_usage_audit = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_usage_audit_period = month {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] internal_service_availability_zone = internal {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] key = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171249] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] live_migration_retry_count = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171426] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_color = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171426] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_config_append = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.171835] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.172600] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_dir = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.172600] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.172600] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_options = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.172840] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_rotate_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.173133] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_rotate_interval_type = days {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.173425] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] log_rotation_type = none {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.173666] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] long_rpc_timeout = 1800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] max_concurrent_builds = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] max_concurrent_live_migrations = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] max_concurrent_snapshots = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] max_local_block_devices = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] max_logfile_count = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] max_logfile_size_mb = 200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] maximum_instance_delete_attempts = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176251] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metadata_listen = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176460] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metadata_listen_port = 8775 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.176693] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metadata_workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.177045] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] migrate_max_retries = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.177349] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] mkisofs_cmd = genisoimage {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.177696] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.177953] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] my_ip = 10.180.1.21 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.178258] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] network_allocate_retries = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.178553] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.178844] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.179142] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] osapi_compute_listen_port = 8774 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.179433] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] osapi_compute_unique_server_name_scope = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.179734] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] osapi_compute_workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.180025] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] password_length = 12 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.180305] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] periodic_enable = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.180599] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] periodic_fuzzy_delay = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.180912] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] pointer_model = usbtablet {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.181227] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] preallocate_images = none {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.181512] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] publish_errors = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.181752] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] pybasedir = /opt/stack/nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.182036] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ram_allocation_ratio = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.182306] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] rate_limit_burst = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.182590] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] rate_limit_except_level = CRITICAL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] rate_limit_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reboot_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reclaim_instance_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] record = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reimage_timeout_per_gb = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] report_interval = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185043] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] rescue_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reserved_host_cpus = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reserved_host_disk_mb = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reserved_host_memory_mb = 512 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] reserved_huge_pages = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] resize_confirm_window = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] resize_fs_using_block_device = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] resume_guests_state_on_host_boot = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185463] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185598] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] rpc_response_timeout = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185755] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] run_external_periodic_tasks = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.185927] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] running_deleted_instance_action = reap {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.186106] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.186274] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] running_deleted_instance_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.186433] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler_instance_sync_interval = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.186600] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_down_time = 720 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.186771] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] servicegroup_driver = db {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.187008] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] shell_completion = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.187195] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] shelved_offload_time = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.187429] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] shelved_poll_interval = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.187705] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] shutdown_timeout = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.187904] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] source_is_ipv6 = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.188082] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ssl_only = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.188335] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.188508] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] sync_power_state_interval = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.188672] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] sync_power_state_pool_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.188842] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] syslog_log_facility = LOG_USER {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.189007] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] tempdir = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.189175] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] timeout_nbd = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.189345] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] transport_url = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.189540] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] update_resources_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.189705] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_cow_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.189865] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_eventlog = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190032] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_journal = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190192] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_json = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190348] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_rootwrap_daemon = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190508] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_stderr = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190661] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] use_syslog = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190819] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vcpu_pin_set = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.190984] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plugging_is_fatal = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.191176] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plugging_timeout = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.191344] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] virt_mkfs = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.191508] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] volume_usage_poll_interval = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.191672] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] watch_log_file = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.191839] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] web = /usr/share/spice-html5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 543.192037] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.192210] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.192375] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.192546] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_concurrency.disable_process_locking = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.193236] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.193433] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.193610] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.193793] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.193971] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.194161] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.194349] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.auth_strategy = keystone {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.194521] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.compute_link_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.194703] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.194882] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.dhcp_domain = novalocal {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.195067] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.enable_instance_password = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.195235] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.glance_link_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.195403] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.195579] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.195746] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.instance_list_per_project_cells = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.195943] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.list_records_by_skipping_down_cells = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.196138] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.local_metadata_per_cell = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.196313] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.max_limit = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.196485] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.metadata_cache_expiration = 15 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.196666] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.neutron_default_tenant_id = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.196840] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.response_validation = warn {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.197093] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.use_neutron_default_nets = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.197287] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.197461] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.197632] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.197808] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.197981] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_dynamic_targets = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.198162] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_jsonfile_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.198345] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.198539] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.backend = dogpile.cache.memcached {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.198711] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.backend_argument = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.198882] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.config_prefix = cache.oslo {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.199061] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.dead_timeout = 60.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.199227] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.debug_cache_backend = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.199391] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.enable_retry_client = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.199576] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.enable_socket_keepalive = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.199774] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.enabled = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.199957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.enforce_fips_mode = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.200133] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.expiration_time = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.200298] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.hashclient_retry_attempts = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.200469] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.200639] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_dead_retry = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.200802] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.200967] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.201144] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.201309] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_pool_maxsize = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.201473] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.201638] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_sasl_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.201817] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.201983] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.202161] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.memcache_username = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.202326] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.proxies = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.202492] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_db = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.202656] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.202830] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.203106] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.203295] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_server = localhost:6379 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.203466] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_socket_timeout = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.203634] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.redis_username = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.203798] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.retry_attempts = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.203964] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.retry_delay = 0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.204146] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.socket_keepalive_count = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.204309] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.socket_keepalive_idle = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.204472] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.socket_keepalive_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.204635] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.tls_allowed_ciphers = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.204794] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.tls_cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.204952] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.tls_certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.205160] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.tls_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.205288] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cache.tls_keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.205457] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.205634] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.205800] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.205974] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.206277] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.206460] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.206630] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.cross_az_attach = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.206798] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.debug = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.206962] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.endpoint_template = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.207214] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.http_retries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.207394] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.207589] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.207781] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.os_region_name = RegionOne {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.207954] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.208133] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cinder.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.208308] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.208583] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.cpu_dedicated_set = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.208770] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.cpu_shared_set = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.208942] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.image_type_exclude_list = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.209123] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.209290] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.209461] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.209641] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.209820] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.209984] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.resource_provider_association_refresh = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.210271] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.210456] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.shutdown_retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.210641] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.210866] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] conductor.workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.210999] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] console.allowed_origins = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.211177] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] console.ssl_ciphers = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.211352] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] console.ssl_minimum_version = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.211523] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] consoleauth.enforce_session_timeout = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.211697] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] consoleauth.token_ttl = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.211895] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.212075] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.212247] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.212409] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.212572] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.212736] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.212901] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.213154] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.213337] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.213500] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.213663] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.213826] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.213987] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.214173] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.service_type = accelerator {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.214337] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.214497] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.214658] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.214817] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.214998] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.215174] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] cyborg.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.215359] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.backend = sqlalchemy {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.215528] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.215696] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.connection_debug = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.215867] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.connection_parameters = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.216038] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.connection_recycle_time = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.216210] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.connection_trace = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.216385] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.db_inc_retry_interval = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.216551] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.db_max_retries = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.216750] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.db_max_retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.216925] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.db_retry_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.217104] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.max_overflow = 50 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.217358] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.max_pool_size = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.217550] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.max_retries = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.217737] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.217902] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.mysql_wsrep_sync_wait = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.218076] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.pool_timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.218244] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.218406] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.slave_connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.218571] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.sqlite_synchronous = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.218735] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] database.use_db_reconnect = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.218917] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.backend = sqlalchemy {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.219095] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.219261] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.connection_debug = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.219432] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.connection_parameters = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.219624] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.connection_recycle_time = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.219795] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.connection_trace = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.219957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.db_inc_retry_interval = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.220132] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.db_max_retries = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.220297] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.db_max_retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.220458] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.db_retry_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.220621] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.max_overflow = 50 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.220780] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.max_pool_size = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.220981] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.max_retries = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.221120] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.221279] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.221435] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.pool_timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.221595] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.retry_interval = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.221756] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.slave_connection = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.221920] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] api_database.sqlite_synchronous = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.222105] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] devices.enabled_mdev_types = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.222288] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.222463] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.222629] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ephemeral_storage_encryption.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.222794] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.222966] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.api_servers = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.223220] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.223407] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.223574] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.223738] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.223900] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.224076] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.debug = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.224245] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.default_trusted_certificate_ids = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.224408] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.enable_certificate_validation = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.224570] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.enable_rbd_download = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.224736] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.224908] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.225088] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.225256] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.225415] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.225580] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.num_retries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.225766] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.rbd_ceph_conf = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.225991] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.rbd_connect_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.226194] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.rbd_pool = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.226368] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.rbd_user = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.226533] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.226695] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.226854] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.227036] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.service_type = image {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.227273] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.227474] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.227643] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.227804] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.227987] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.228171] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.verify_glance_signatures = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.228333] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] glance.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.228502] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] guestfs.debug = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.228671] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] mks.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.229040] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.229237] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] image_cache.manager_interval = 2400 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.229409] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] image_cache.precache_concurrency = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.229604] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] image_cache.remove_unused_base_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.229784] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.229957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.230151] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] image_cache.subdirectory_name = _base {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.230330] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.api_max_retries = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.230501] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.api_retry_interval = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.230663] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.230827] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.230989] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.231166] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.231332] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.231499] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.conductor_group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.231661] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.231823] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.231983] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.232159] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.232320] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.232482] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.232641] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.232806] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.peer_list = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.232966] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.233146] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.233408] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.serial_console_state_timeout = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.233587] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.233763] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.service_type = baremetal {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.233926] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.shard = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.234104] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.234269] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.234430] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.234590] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.234773] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.234936] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ironic.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.235132] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.235312] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] key_manager.fixed_key = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.235498] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.235665] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.barbican_api_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.235827] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.barbican_endpoint = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236007] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.barbican_endpoint_type = public {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236180] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.barbican_region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236341] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236507] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236672] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236832] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.236991] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.237168] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.number_of_retries = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.237419] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.retry_delay = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.237622] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.send_service_user_token = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.237796] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.237958] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.238137] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.verify_ssl = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.238298] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican.verify_ssl_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.238467] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.238633] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.238795] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.238957] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.239139] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.239306] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.239495] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.239671] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.239836] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] barbican_service_user.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.240015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.approle_role_id = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.240184] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.approle_secret_id = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.240357] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.kv_mountpoint = secret {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.240554] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.kv_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.240730] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.kv_version = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.240897] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.namespace = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.241070] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.root_token_id = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.241238] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.ssl_ca_crt_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.241410] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.timeout = 60.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.241575] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.use_ssl = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.241749] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.241921] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.242097] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.242264] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.242427] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.242593] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.242755] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.242915] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.243089] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.243256] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.243502] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.243679] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.243840] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244009] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244178] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244337] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244509] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.service_type = identity {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244674] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244834] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.244993] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.245166] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.245348] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.245512] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] keystone.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.245717] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.connection_uri = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.245879] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_mode = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.246055] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.246229] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_models = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.246404] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_power_governor_high = performance {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.246576] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.246743] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_power_management = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.246915] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.247096] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.device_detach_attempts = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.247265] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.device_detach_timeout = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.247528] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.disk_cachemodes = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.247722] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.disk_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.247895] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.enabled_perf_events = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.248075] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.file_backed_memory = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.248250] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.gid_maps = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.248412] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.hw_disk_discard = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.248569] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.hw_machine_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.248746] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_rbd_ceph_conf = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.248919] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.249095] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.249272] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_rbd_glance_store_name = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.249455] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_rbd_pool = rbd {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.249654] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_type = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.249825] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.images_volume_group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.249992] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.inject_key = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.250210] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.inject_partition = -2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.250383] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.inject_password = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.250551] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.iscsi_iface = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.250718] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.iser_use_multipath = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.250885] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.251062] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.251230] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_downtime = 500 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.251396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.251559] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.251723] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_inbound_addr = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.251888] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.252066] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.252235] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_scheme = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.252413] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_timeout_action = abort {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.252578] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_tunnelled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.252742] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_uri = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.252905] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.live_migration_with_native_tls = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.253077] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.max_queues = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.253244] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.253561] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.253752] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.nfs_mount_options = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.254071] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.254254] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.254423] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.254586] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.254753] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.254916] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.num_pcie_ports = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.255096] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.255269] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.pmem_namespaces = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.255432] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.quobyte_client_cfg = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.255737] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.255917] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.256098] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.256268] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.256432] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rbd_secret_uuid = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.256594] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rbd_user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.256766] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.256943] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.257121] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rescue_image_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.257283] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rescue_kernel_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.257460] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rescue_ramdisk_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.257721] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.257893] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.rx_queue_size = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.258079] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.smbfs_mount_options = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.258363] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.258539] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.snapshot_compression = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.258708] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.snapshot_image_format = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.258929] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.259109] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.sparse_logical_volumes = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.259277] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.swtpm_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.259450] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.swtpm_group = tss {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.259650] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.swtpm_user = tss {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.259826] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.sysinfo_serial = unique {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.259988] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.tb_cache_size = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.260166] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.tx_queue_size = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.260332] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.uid_maps = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.260510] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.use_virtio_for_bridges = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.260697] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.virt_type = kvm {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.260871] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.volume_clear = zero {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.261048] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.volume_clear_size = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.261220] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.volume_use_multipath = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.261425] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_cache_path = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.261550] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.261722] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.261922] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.262120] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.262402] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.262605] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.vzstorage_mount_user = stack {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.262790] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.262967] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.263159] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.263325] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.263486] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.263740] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.263911] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.264082] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.264258] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.default_floating_pool = public {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.264420] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.264582] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.extension_sync_interval = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.264747] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.http_retries = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.264908] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.265079] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.265244] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.265418] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.265579] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.265777] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.ovs_bridge = br-int {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.265914] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.physnets = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.266097] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.region_name = RegionOne {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.266265] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.266436] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.service_metadata_proxy = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.266600] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.266774] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.service_type = network {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.266937] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.267110] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.267269] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.267428] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.267674] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.267863] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] neutron.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.268047] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] notifications.bdms_in_notifications = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.268234] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] notifications.default_level = INFO {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.268413] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] notifications.notification_format = unversioned {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.268580] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] notifications.notify_on_state_change = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.268762] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.268940] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] pci.alias = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.269125] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] pci.device_spec = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.269293] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] pci.report_in_placement = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.269484] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.269678] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.269854] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.270030] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.270201] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.270367] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.270551] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.270727] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.270891] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.default_domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.271062] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.default_domain_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.271226] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.271383] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.domain_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.271540] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.271700] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.271883] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.272063] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.272225] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.272396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.272557] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.project_domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.272727] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.project_domain_name = Default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.272898] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.project_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.273138] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.project_name = service {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.273260] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.region_name = RegionOne {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.273426] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.273590] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.273867] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.service_type = placement {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.274073] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.274244] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.274410] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.274595] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.system_scope = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.274776] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.274938] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.trust_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.275110] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.user_domain_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.275282] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.user_domain_name = Default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.275443] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.user_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.275616] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.username = nova {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.275801] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.275967] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] placement.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.276157] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.cores = 20 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.276324] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.count_usage_from_placement = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.276500] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.276678] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.injected_file_content_bytes = 10240 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.276844] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.injected_file_path_length = 255 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.277021] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.injected_files = 5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.277188] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.instances = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.277354] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.key_pairs = 100 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.277525] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.metadata_items = 128 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.277754] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.ram = 51200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.277946] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.recheck_quota = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.278131] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.server_group_members = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.278301] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] quota.server_groups = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.278473] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.278642] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.278808] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.image_metadata_prefilter = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.278969] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.279150] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.max_attempts = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.279322] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.max_placement_results = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.279513] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.279685] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.279848] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.280032] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] scheduler.workers = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.280213] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.280386] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.280596] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.280774] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.280941] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.281122] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.281293] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.281489] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.281665] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.host_subset_size = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.281857] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.282049] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.282225] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.282396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.isolated_hosts = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.282561] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.isolated_images = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.282728] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.282889] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.283064] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.283243] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.pci_in_placement = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.283409] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.283574] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.283818] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.284014] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.284191] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.284360] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.284527] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.track_instance_changes = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.284707] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.284879] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metrics.required = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.285055] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metrics.weight_multiplier = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.285228] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.285396] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] metrics.weight_setting = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.285720] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.285919] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] serial_console.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.286137] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] serial_console.port_range = 10000:20000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.286319] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.286494] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.286691] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] serial_console.serialproxy_port = 6083 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.286871] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.287069] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.auth_type = password {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.287238] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.287400] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.287563] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.287728] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.287969] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.288163] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.send_service_user_token = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.288332] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.288507] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] service_user.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.288683] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.agent_enabled = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.288849] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.289171] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.289364] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.289561] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.html5proxy_port = 6082 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.289737] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.image_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.289900] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.jpeg_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.290071] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.playback_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.290237] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.require_secure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.290410] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.server_listen = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.290610] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.290775] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.streaming_mode = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.290935] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] spice.zlib_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.291116] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] upgrade_levels.baseapi = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.291291] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] upgrade_levels.compute = auto {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.291453] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] upgrade_levels.conductor = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.291612] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] upgrade_levels.scheduler = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.291789] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.291973] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.292151] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.292314] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.292478] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.292643] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.292802] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.292965] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.293135] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vendordata_dynamic_auth.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.293311] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.api_retry_count = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.293476] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.ca_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.293650] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.293820] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.cluster_name = testcl1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.294090] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.connection_pool_size = 10 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.294264] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.console_delay_seconds = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.294440] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.datastore_regex = ^datastore.* {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.294643] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.294822] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.host_password = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.294995] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.host_port = 443 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.295182] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.host_username = administrator@vsphere.local {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.295355] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.insecure = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.295521] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.integration_bridge = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.295689] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.maximum_objects = 100 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.295850] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.pbm_default_policy = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.296060] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.pbm_enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.296187] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.pbm_wsdl_location = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.296356] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.296517] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.serial_port_proxy_uri = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.296679] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.serial_port_service_uri = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.296843] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.task_poll_interval = 0.5 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.297025] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.use_linked_clone = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.297204] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.vnc_keymap = en-us {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.297375] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.vnc_port = 5900 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.297539] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vmware.vnc_port_total = 10000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.297727] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.auth_schemes = ['none'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.297990] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.298321] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.298517] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.298718] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.novncproxy_port = 6080 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.298912] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.server_listen = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.299104] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.299274] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.vencrypt_ca_certs = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.299437] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.vencrypt_client_cert = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.299627] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vnc.vencrypt_client_key = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.299814] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.299979] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.disable_deep_image_inspection = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.300156] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.300345] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.300487] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.300682] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.disable_rootwrap = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.300847] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.enable_numa_live_migration = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301015] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301187] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301349] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301511] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.libvirt_disable_apic = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301673] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301861] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.301997] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.302172] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.302334] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.302495] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.302660] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.302819] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.302978] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.303156] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.303340] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.303509] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.client_socket_timeout = 900 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.303679] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.default_pool_size = 1000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.303847] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.keep_alive = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.304101] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.max_header_line = 16384 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.304287] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.304456] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.ssl_ca_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.304620] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.ssl_cert_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.304787] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.ssl_key_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.304954] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.tcp_keepidle = 600 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.305149] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.305322] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] zvm.ca_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.305481] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] zvm.cloud_connector_url = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.305767] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.305941] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] zvm.reachable_timeout = 300 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.306155] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.enforce_new_defaults = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.306520] env[62952]: WARNING oslo_config.cfg [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 543.306708] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.enforce_scope = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.306944] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.policy_default_rule = default {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.307168] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.307355] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.policy_file = policy.yaml {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.307532] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.307699] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.307860] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.308099] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.308286] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.308459] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.308639] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.308822] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.connection_string = messaging:// {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.308991] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.enabled = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.309181] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.es_doc_type = notification {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.309347] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.es_scroll_size = 10000 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.309540] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.es_scroll_time = 2m {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.309713] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.filter_error_trace = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.309885] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.hmac_keys = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.310067] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.sentinel_service_name = mymaster {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.310239] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.socket_timeout = 0.1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.310405] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.trace_requests = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.310593] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler.trace_sqlalchemy = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.310789] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler_jaeger.process_tags = {} {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.310954] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler_jaeger.service_name_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.311134] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] profiler_otlp.service_name_prefix = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.311299] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] remote_debug.host = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.311460] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] remote_debug.port = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.311641] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.311804] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.311972] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.312143] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.312306] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.312466] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.312629] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.312791] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.312950] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.313135] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.313297] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.313467] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.313634] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.313803] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.313974] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.314232] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.314409] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.314586] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.314844] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.315041] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.315217] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.315385] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.315549] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.315720] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.315883] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.316055] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.316254] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.316382] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.316549] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.316717] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.ssl = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.316895] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.317074] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.317348] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.317549] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.317729] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.317893] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.318153] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.318343] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_notifications.retry = -1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.318536] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.318800] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.318979] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.auth_section = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.319161] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.auth_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.319323] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.cafile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.319498] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.certfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.319681] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.collect_timing = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.319842] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.connect_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.320008] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.connect_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.320175] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.endpoint_id = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.320335] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.endpoint_override = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.320522] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.insecure = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.320702] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.keyfile = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.320864] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.max_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.321032] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.min_version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.321197] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.region_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.321359] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.retriable_status_codes = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.321519] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.service_name = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.321680] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.service_type = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.321865] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.split_loggers = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.322079] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.status_code_retries = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.322210] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.status_code_retry_delay = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.322368] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.timeout = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.322527] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.valid_interfaces = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.322685] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_limit.version = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.322851] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_reports.file_event_handler = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.323024] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.323189] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] oslo_reports.log_dir = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.323369] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.323536] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.323695] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.323895] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.324079] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.324328] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.324510] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.324677] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_ovs_privileged.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.324839] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.325019] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.325183] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.325344] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] vif_plug_ovs_privileged.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.325516] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.325700] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.325878] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.326063] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.326243] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.326412] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.326580] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.326746] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.326926] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.327167] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.isolate_vif = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.327361] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.327537] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.327710] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.327882] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.328057] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] os_vif_ovs.per_port_bridge = False {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.328315] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] privsep_osbrick.capabilities = [21] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.328489] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] privsep_osbrick.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.328654] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] privsep_osbrick.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.328822] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.328987] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.329162] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] privsep_osbrick.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.329337] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.329525] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] nova_sys_admin.group = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.329696] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] nova_sys_admin.helper_command = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.329865] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.330039] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.330204] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] nova_sys_admin.user = None {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 543.330334] env[62952]: DEBUG oslo_service.service [None req-ba8c6239-2b8d-4581-8342-a10208fae8c2 None None] ******************************************************************************** {{(pid=62952) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 543.330844] env[62952]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 543.834176] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Getting list of instances from cluster (obj){ [ 543.834176] env[62952]: value = "domain-c8" [ 543.834176] env[62952]: _type = "ClusterComputeResource" [ 543.834176] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 543.835350] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51b4b04-aaf4-48ec-8089-e9c4d23b32e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.844479] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Got total of 0 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 543.845164] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 543.845638] env[62952]: INFO nova.virt.node [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Generated node identity 9b9257d4-3be7-42a5-ae8c-67cf64823681 [ 543.845873] env[62952]: INFO nova.virt.node [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Wrote node identity 9b9257d4-3be7-42a5-ae8c-67cf64823681 to /opt/stack/data/n-cpu-1/compute_id [ 544.348197] env[62952]: WARNING nova.compute.manager [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Compute nodes ['9b9257d4-3be7-42a5-ae8c-67cf64823681'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 545.354145] env[62952]: INFO nova.compute.manager [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 546.362020] env[62952]: WARNING nova.compute.manager [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 546.362020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.362020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.362020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.362020] env[62952]: DEBUG nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 546.362020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487707fb-3f5c-43b8-bdbf-61efcb9d72f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.370443] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a92f410-7f79-40d9-aff1-1bfa2ae19d0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.384460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed4b9fa-46f2-4768-96b0-8a40b3801c3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.391483] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67400826-656c-4687-9976-036e50ed0132 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.422092] env[62952]: DEBUG nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181505MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 546.422484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.422794] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.927378] env[62952]: WARNING nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] No compute node record for cpu-1:9b9257d4-3be7-42a5-ae8c-67cf64823681: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 9b9257d4-3be7-42a5-ae8c-67cf64823681 could not be found. [ 547.430781] env[62952]: INFO nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 9b9257d4-3be7-42a5-ae8c-67cf64823681 [ 548.941742] env[62952]: DEBUG nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 548.941742] env[62952]: DEBUG nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 549.091000] env[62952]: INFO nova.scheduler.client.report [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] [req-d430709f-faa8-40d7-bace-a8830fb459bd] Created resource provider record via placement API for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 549.105974] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a25d76-5b9d-43b3-b129-7964969d72a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.114077] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1004d9-cd7c-4446-b2d9-229e27e88062 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.143703] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa7c2aa-64ad-4d48-9418-cb926f802148 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.152023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099bb220-212e-4b5b-9497-ba12b46bf75f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.163835] env[62952]: DEBUG nova.compute.provider_tree [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 549.699772] env[62952]: DEBUG nova.scheduler.client.report [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 549.700118] env[62952]: DEBUG nova.compute.provider_tree [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 0 to 1 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 549.700341] env[62952]: DEBUG nova.compute.provider_tree [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 549.753285] env[62952]: DEBUG nova.compute.provider_tree [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 1 to 2 during operation: update_traits {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 550.257730] env[62952]: DEBUG nova.compute.resource_tracker [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 550.258178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.835s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.258178] env[62952]: DEBUG nova.service [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Creating RPC server for service compute {{(pid=62952) start /opt/stack/nova/nova/service.py:186}} [ 550.272160] env[62952]: DEBUG nova.service [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] Join ServiceGroup membership for this service compute {{(pid=62952) start /opt/stack/nova/nova/service.py:203}} [ 550.272356] env[62952]: DEBUG nova.servicegroup.drivers.db [None req-aabd6bf7-66b4-4b13-9c56-5715524ef4c8 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62952) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 558.274690] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.777871] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Getting list of instances from cluster (obj){ [ 558.777871] env[62952]: value = "domain-c8" [ 558.777871] env[62952]: _type = "ClusterComputeResource" [ 558.777871] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 558.779121] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13915ab6-3dc1-471d-829c-5fe263b2dad3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.787917] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Got total of 0 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 558.788162] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.788456] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Getting list of instances from cluster (obj){ [ 558.788456] env[62952]: value = "domain-c8" [ 558.788456] env[62952]: _type = "ClusterComputeResource" [ 558.788456] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 558.789277] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6409c52-3c2d-44b5-9908-e1a526f58624 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.796233] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Got total of 0 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 587.271271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "977675cb-bcf3-431b-aef1-36f8232b093b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.271582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "977675cb-bcf3-431b-aef1-36f8232b093b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.774347] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.892620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "418cd947-00b8-485f-b9fc-2a675c58998d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.892620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "418cd947-00b8-485f-b9fc-2a675c58998d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.323751] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.324104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.325929] env[62952]: INFO nova.compute.claims [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.369734] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquiring lock "e075df95-2102-4f64-b4ba-a65b8fea78d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.371651] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Lock "e075df95-2102-4f64-b4ba-a65b8fea78d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.394943] env[62952]: DEBUG nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.872190] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.925273] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.406218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.419203] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85c4fff-7db2-4981-8462-7afbd2e2a2a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.427573] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f1daac-df4d-4149-9729-b6eac60b8c44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.463876] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5202a7fd-f88b-4644-a03d-60ed1a2487d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.471999] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a12555-e687-425a-8399-409df466595c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.488965] env[62952]: DEBUG nova.compute.provider_tree [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.992276] env[62952]: DEBUG nova.scheduler.client.report [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.263379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "7f376384-c401-418e-94d3-c9136005a11e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.263801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "7f376384-c401-418e-94d3-c9136005a11e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.502019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.175s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.502019] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.503178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.578s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.507134] env[62952]: INFO nova.compute.claims [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.766714] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.014151] env[62952]: DEBUG nova.compute.utils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.014878] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.015343] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.290803] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.523413] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.656824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e00301-f60a-41db-a74e-6132f20e09d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.666528] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bd78a8-25bf-4817-9c92-bdcc8cacb0f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.709929] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b054d7ef-ca9f-406c-9681-d3a4c162df69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.720485] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0c5bc6-f5e3-4460-b24c-1c5d9f7d7ba4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.740598] env[62952]: DEBUG nova.compute.provider_tree [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.936034] env[62952]: DEBUG nova.policy [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ae217ec66fe45beb1381449073980d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23e42f4ddc5a4d96b3682876928f1dfc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.246721] env[62952]: DEBUG nova.scheduler.client.report [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.545071] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.581635] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.581922] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.583646] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.584618] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.584618] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.584618] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.584618] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.584618] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.585026] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.585100] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.585669] env[62952]: DEBUG nova.virt.hardware [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.586323] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0423ec0a-b2ad-4c99-b567-733869bcafa1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.596864] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cbbf67-5ee2-4482-a20a-3d51829dd62a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.614664] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9486698-5bfb-42b3-9340-e58fe71be1a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.753137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.753546] env[62952]: DEBUG nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.756931] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.350s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.759370] env[62952]: INFO nova.compute.claims [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.264355] env[62952]: DEBUG nova.compute.utils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.269124] env[62952]: DEBUG nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 593.458100] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Successfully created port: bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.769728] env[62952]: DEBUG nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.879390] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27b91de-0c98-4f5a-8a09-52ea9ff71d8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.888731] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f1f46c-2b31-4c0f-b978-8623e81601d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.927543] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7833d707-8de5-4a97-b8b0-30c806a4ab31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.937609] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5a0f5c-576f-4e30-b449-e00175fb2b82 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.953417] env[62952]: DEBUG nova.compute.provider_tree [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.457703] env[62952]: DEBUG nova.scheduler.client.report [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.784782] env[62952]: DEBUG nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.821746] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.821990] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.822368] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.822720] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.822931] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.823140] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.823319] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.823468] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.823630] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.823788] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.823985] env[62952]: DEBUG nova.virt.hardware [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.825231] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf7d3ca-fe14-47a3-bfdf-f953051345bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.835281] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614aae87-4118-4166-9ccc-8131f94927e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.851197] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 594.861913] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 594.861913] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b378e4af-c962-4de6-846d-cfeeb764f166 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.874309] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Created folder: OpenStack in parent group-v4. [ 594.874464] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating folder: Project (abbfb9b0c5ba4f1cab97b1b6bd0fa439). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 594.874867] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa55ad41-7f03-4fbd-b60e-ab35e6fbf55d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.885791] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Created folder: Project (abbfb9b0c5ba4f1cab97b1b6bd0fa439) in parent group-v290852. [ 594.885884] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating folder: Instances. Parent ref: group-v290853. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 594.886583] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b894d300-d906-434b-9322-f13a7011cde5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.896683] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Created folder: Instances in parent group-v290853. [ 594.896988] env[62952]: DEBUG oslo.service.loopingcall [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.897212] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 594.897661] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa4528a0-9363-44bd-86f2-b2e94c7cc9fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.915349] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 594.915349] env[62952]: value = "task-1367018" [ 594.915349] env[62952]: _type = "Task" [ 594.915349] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.926529] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367018, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.971590] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.975018] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 594.978864] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.688s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.980026] env[62952]: INFO nova.compute.claims [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.430900] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367018, 'name': CreateVM_Task, 'duration_secs': 0.274734} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.431090] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 595.433656] env[62952]: DEBUG oslo_vmware.service [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef520f79-40a3-4dea-b8b8-386630612a36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.441610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.442239] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.443379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 595.446291] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67a88c26-070b-4669-bcc2-7e551f496b4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.450323] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 595.450323] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bb2117-d06c-6b71-a639-f4e8a57193df" [ 595.450323] env[62952]: _type = "Task" [ 595.450323] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.460863] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52bb2117-d06c-6b71-a639-f4e8a57193df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.479477] env[62952]: DEBUG nova.compute.utils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.484253] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.484253] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 595.815274] env[62952]: DEBUG nova.policy [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3194a809559f4730815775cef88e849c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8eb7cde31904e13b7e83bcf59a5106c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.969015] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.969323] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 595.969554] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.969641] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.970116] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 595.970373] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06b6fb67-cef6-48bf-819c-83241a04b122 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.985226] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.990963] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 595.991168] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 595.991938] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a42433-37fb-4f3d-ac19-2914aa114571 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.002549] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-433fe10a-320d-4468-b936-a1308cd17d5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.009506] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 596.009506] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5250619b-05f5-42d0-bf1a-0324794d3cd1" [ 596.009506] env[62952]: _type = "Task" [ 596.009506] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.022604] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5250619b-05f5-42d0-bf1a-0324794d3cd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.108340] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89de8f68-4a13-4cbf-99e0-49a9fc973359 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.120827] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82864be1-af73-4b89-9ddf-b50bcd5a8276 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.156622] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4a5311-1802-41da-ba1f-add5028de6d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.165375] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11eaa551-8baf-4735-88e3-817b6b1dc390 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.186082] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.521468] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 596.521867] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating directory with path [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 596.522230] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96728dab-d026-41f2-8504-16457f732fa8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.551025] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Created directory with path [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 596.551025] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Fetch image to [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 596.551025] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Downloading image file data e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk on the data store datastore2 {{(pid=62952) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 596.551025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4498570f-c8e2-453a-9796-99095bd350d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.558612] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8fbc8e-f37b-4803-8ba2-06a4acb6e2c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.575307] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40274a3-0b7f-4121-8564-0ea14d6cd9aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.616601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609e40ff-0885-4313-8cdc-6538ce4b5b8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.624249] env[62952]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a087498a-8b59-4ecd-8ecc-9fc97e560418 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.658537] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Downloading image file data e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to the data store datastore2 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 596.694093] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.765069] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 597.002305] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.036681] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.036681] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.038734] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.038734] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.038734] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.038908] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.040012] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.040012] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.040012] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.040012] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.040012] env[62952]: DEBUG nova.virt.hardware [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.043094] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2485021-0480-4c15-bda5-2981e790c872 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.061378] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1d3ed1-056d-4419-bdca-14c8b37a3be8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.166109] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "77418633-2833-47a7-b30f-dca84432d382" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.166243] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "77418633-2833-47a7-b30f-dca84432d382" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.208520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.208520] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.222630] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Successfully created port: 70ce59a5-bba1-4503-950e-c3bffab25105 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.512277] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 597.513051] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 597.656023] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Downloaded image file data e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk on the data store datastore2 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 597.657444] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 597.658372] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Copying Virtual Disk [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk to [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 597.658987] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6479463b-eba9-4cc9-a70d-8e24b4ce7312 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.671619] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.675830] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 597.675830] env[62952]: value = "task-1367019" [ 597.675830] env[62952]: _type = "Task" [ 597.675830] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.685018] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367019, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.712417] env[62952]: DEBUG nova.compute.utils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.714729] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.714729] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 597.961060] env[62952]: DEBUG nova.policy [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84a933f0beb740bb96a0ffab4c44568b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c6d16afcdb245ff97759f9b6ed1ed43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.966905] env[62952]: ERROR nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 597.966905] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.966905] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.966905] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.966905] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.966905] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.966905] env[62952]: ERROR nova.compute.manager raise self.value [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.966905] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.966905] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.966905] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.967570] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.967570] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.967570] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 597.967570] env[62952]: ERROR nova.compute.manager [ 597.967721] env[62952]: Traceback (most recent call last): [ 597.967765] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.967765] env[62952]: listener.cb(fileno) [ 597.967765] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.967765] env[62952]: result = function(*args, **kwargs) [ 597.967765] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.967765] env[62952]: return func(*args, **kwargs) [ 597.967765] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.967765] env[62952]: raise e [ 597.967765] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.967765] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 597.967765] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.967765] env[62952]: created_port_ids = self._update_ports_for_instance( [ 597.967765] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.967765] env[62952]: with excutils.save_and_reraise_exception(): [ 597.967765] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.967765] env[62952]: self.force_reraise() [ 597.967765] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.967765] env[62952]: raise self.value [ 597.967765] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.967765] env[62952]: updated_port = self._update_port( [ 597.967765] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.967765] env[62952]: _ensure_no_port_binding_failure(port) [ 597.968392] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.968392] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.968392] env[62952]: nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 597.968392] env[62952]: Removing descriptor: 15 [ 597.970466] env[62952]: ERROR nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Traceback (most recent call last): [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] yield resources [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self.driver.spawn(context, instance, image_meta, [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] vm_ref = self.build_virtual_machine(instance, [ 597.970466] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] for vif in network_info: [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return self._sync_wrapper(fn, *args, **kwargs) [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self.wait() [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self[:] = self._gt.wait() [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return self._exit_event.wait() [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.970772] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] result = hub.switch() [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return self.greenlet.switch() [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] result = function(*args, **kwargs) [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return func(*args, **kwargs) [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] raise e [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] nwinfo = self.network_api.allocate_for_instance( [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] created_port_ids = self._update_ports_for_instance( [ 597.971092] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] with excutils.save_and_reraise_exception(): [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self.force_reraise() [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] raise self.value [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] updated_port = self._update_port( [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] _ensure_no_port_binding_failure(port) [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] raise exception.PortBindingFailed(port_id=port['id']) [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 597.971408] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] [ 597.971777] env[62952]: INFO nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Terminating instance [ 597.974359] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.974521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquired lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.974697] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.190993] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367019, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.201875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.202074] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.203708] env[62952]: INFO nova.compute.claims [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.218543] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.538166] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.555109] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquiring lock "21c8106e-8b6c-447b-89e2-df1b34aed12c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.555450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Lock "21c8106e-8b6c-447b-89e2-df1b34aed12c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.687183] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367019, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670673} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.687623] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Copied Virtual Disk [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk to [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 598.687836] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleting the datastore file [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 598.688925] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.689999] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d92b12d-5462-4d3e-88be-4e9e82cc4abc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.697836] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 598.697836] env[62952]: value = "task-1367020" [ 598.697836] env[62952]: _type = "Task" [ 598.697836] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.713075] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.722466] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 598.724318] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 598.724860] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 598.724860] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 599.060199] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.197331] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Releasing lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.197331] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.197331] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.197331] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3437b4e2-1818-487e-bfb6-c9e0ac7a66c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.213542] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641a77d1-3721-41b3-a307-9cc224b26557 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.231710] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Successfully created port: 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.243746] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.243904] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.244048] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.244205] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.244332] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 77418633-2833-47a7-b30f-dca84432d382] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 599.244459] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 599.245483] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.248912] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.253822] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.254208] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02452} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.255523] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.256064] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 599.256268] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Moving file from [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff/e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77. {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 599.257276] env[62952]: DEBUG nova.compute.manager [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Received event network-changed-bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.257694] env[62952]: DEBUG nova.compute.manager [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Refreshing instance network info cache due to event network-changed-bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 599.257913] env[62952]: DEBUG oslo_concurrency.lockutils [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] Acquiring lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.258063] env[62952]: DEBUG oslo_concurrency.lockutils [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] Acquired lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.258235] env[62952]: DEBUG nova.network.neutron [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Refreshing network info cache for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 599.266904] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.267759] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-68020d81-8d7c-41c7-a929-eaeb7801d086 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.270491] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 977675cb-bcf3-431b-aef1-36f8232b093b could not be found. [ 599.270728] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.271101] env[62952]: INFO nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Took 0.08 seconds to destroy the instance on the hypervisor. [ 599.271338] env[62952]: DEBUG oslo.service.loopingcall [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.271829] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.272759] env[62952]: DEBUG nova.compute.manager [-] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.273027] env[62952]: DEBUG nova.network.neutron [-] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.274874] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.275080] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 599.278144] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.282528] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 599.282528] env[62952]: value = "task-1367021" [ 599.282528] env[62952]: _type = "Task" [ 599.282528] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.294939] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367021, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.298635] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.300269] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.300269] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.300269] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.300269] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.300269] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.300269] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.300644] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.300644] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.300644] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.300770] env[62952]: DEBUG nova.virt.hardware [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.302330] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6baa077-ff04-442e-abd2-bab68d235d1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.318092] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94589bad-a265-492a-ae0b-e8a31f511337 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.321109] env[62952]: DEBUG nova.network.neutron [-] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.432020] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a20251-67a6-4d63-9f92-0423e99f7223 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.439830] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d242612-aeef-4034-b8c9-099f21922684 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.481382] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689e6c5b-8a85-4954-a6f3-d52dec97a81c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.490318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11972b7-c96a-447d-ace2-d86cebd6e2fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.516724] env[62952]: DEBUG nova.compute.provider_tree [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.600819] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.786252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.795511] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367021, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026091} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.795743] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] File moved {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 599.803028] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Cleaning up location [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 599.803028] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleting the datastore file [datastore2] vmware_temp/2d6aa63b-1147-4cda-819c-f9aad472e6ff {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 599.803028] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b07f1c56-cf0e-470f-b2a7-b0babdf42bd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.807510] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 599.807510] env[62952]: value = "task-1367022" [ 599.807510] env[62952]: _type = "Task" [ 599.807510] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.818192] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.823604] env[62952]: DEBUG nova.network.neutron [-] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.895506] env[62952]: DEBUG nova.network.neutron [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.023933] env[62952]: DEBUG nova.scheduler.client.report [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.170879] env[62952]: DEBUG nova.network.neutron [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.323860] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367022, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025552} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.324130] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 600.325974] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af90201a-61ee-4462-a0c4-105517caece7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.331668] env[62952]: INFO nova.compute.manager [-] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Took 1.06 seconds to deallocate network for instance. [ 600.336275] env[62952]: DEBUG nova.compute.claims [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.336275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.338098] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 600.338098] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52032e10-561c-e724-ab7a-85cd63bb2521" [ 600.338098] env[62952]: _type = "Task" [ 600.338098] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.348112] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52032e10-561c-e724-ab7a-85cd63bb2521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.529377] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.532246] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.533719] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.932s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.535292] env[62952]: INFO nova.compute.claims [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.673934] env[62952]: DEBUG oslo_concurrency.lockutils [req-4936df45-167d-4147-8c93-b7dbfbc9e0d7 req-b45d89d0-759e-40f7-93f1-817f6526ddc4 service nova] Releasing lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.853825] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52032e10-561c-e724-ab7a-85cd63bb2521, 'name': SearchDatastore_Task, 'duration_secs': 0.008715} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.862021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.862021] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 600.862021] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ccdc4150-9d46-41ae-877b-65ee7fa9b8e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.871371] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 600.871371] env[62952]: value = "task-1367023" [ 600.871371] env[62952]: _type = "Task" [ 600.871371] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.882533] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.044558] env[62952]: DEBUG nova.compute.utils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.052188] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 601.052188] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 601.383794] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367023, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.400424] env[62952]: DEBUG nova.policy [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4e5d87441d1430f84b10a8397b4834e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9908ca9c44c41c7a0cfe9e1af34a4aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.549404] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.717219] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67984270-c3e4-4a12-90a6-672bab8f2410 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.725901] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc10112a-633a-48d0-bd2c-3cffb241fcfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.761131] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56825106-7213-41db-baef-6b7ae5e3feb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.768601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8e1bd0-847f-4690-886f-6a3f954fd4ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.782721] env[62952]: DEBUG nova.compute.provider_tree [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 601.882374] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557555} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.882667] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 601.882921] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 601.883338] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-333227cc-6662-4ce9-9714-f39b17cee1d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.889684] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 601.889684] env[62952]: value = "task-1367024" [ 601.889684] env[62952]: _type = "Task" [ 601.889684] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.897076] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367024, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.156078] env[62952]: ERROR nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 602.156078] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.156078] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.156078] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.156078] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.156078] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.156078] env[62952]: ERROR nova.compute.manager raise self.value [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.156078] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.156078] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.156078] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.156511] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.156511] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.156511] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 602.156511] env[62952]: ERROR nova.compute.manager [ 602.156511] env[62952]: Traceback (most recent call last): [ 602.156511] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.156511] env[62952]: listener.cb(fileno) [ 602.156511] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.156511] env[62952]: result = function(*args, **kwargs) [ 602.156511] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.156511] env[62952]: return func(*args, **kwargs) [ 602.156511] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.156511] env[62952]: raise e [ 602.156511] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.156511] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 602.156511] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.156511] env[62952]: created_port_ids = self._update_ports_for_instance( [ 602.156511] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.156511] env[62952]: with excutils.save_and_reraise_exception(): [ 602.156511] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.156511] env[62952]: self.force_reraise() [ 602.156511] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.156511] env[62952]: raise self.value [ 602.156511] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.156511] env[62952]: updated_port = self._update_port( [ 602.156511] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.156511] env[62952]: _ensure_no_port_binding_failure(port) [ 602.156511] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.156511] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.157231] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 602.157231] env[62952]: Removing descriptor: 17 [ 602.157231] env[62952]: ERROR nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Traceback (most recent call last): [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] yield resources [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self.driver.spawn(context, instance, image_meta, [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.157231] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] vm_ref = self.build_virtual_machine(instance, [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] for vif in network_info: [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return self._sync_wrapper(fn, *args, **kwargs) [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self.wait() [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self[:] = self._gt.wait() [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return self._exit_event.wait() [ 602.157513] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] result = hub.switch() [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return self.greenlet.switch() [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] result = function(*args, **kwargs) [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return func(*args, **kwargs) [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] raise e [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] nwinfo = self.network_api.allocate_for_instance( [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.157845] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] created_port_ids = self._update_ports_for_instance( [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] with excutils.save_and_reraise_exception(): [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self.force_reraise() [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] raise self.value [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] updated_port = self._update_port( [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] _ensure_no_port_binding_failure(port) [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.160015] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] raise exception.PortBindingFailed(port_id=port['id']) [ 602.160425] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 602.160425] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] [ 602.160425] env[62952]: INFO nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Terminating instance [ 602.160425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquiring lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.160425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquired lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.160425] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.321475] env[62952]: ERROR nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [req-baa62556-8e1c-4d49-9b46-ac2d6a057406] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-baa62556-8e1c-4d49-9b46-ac2d6a057406"}]} [ 602.340475] env[62952]: DEBUG nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 602.357167] env[62952]: DEBUG nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 602.357581] env[62952]: DEBUG nova.compute.provider_tree [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 602.371459] env[62952]: DEBUG nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 602.405238] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367024, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.405881] env[62952]: DEBUG nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 602.408708] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 602.411203] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87cb017-8063-4e2c-ac10-ce6eb6ba3c69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.435034] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 602.438949] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-028bae96-8697-4500-a52f-10f301eda2ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.461070] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 602.461070] env[62952]: value = "task-1367025" [ 602.461070] env[62952]: _type = "Task" [ 602.461070] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.473605] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367025, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.543049] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b72d4f-a28b-452b-8935-6d21ecab28fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.550601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467929a4-0f1a-469b-90ef-2cac8a3a15ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.593019] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.598443] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6dbeea-a05a-4b5c-b81c-b2578b9f6a0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.608914] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3db835-3d1a-48b5-92ff-5b9b914e95f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.626842] env[62952]: DEBUG nova.compute.provider_tree [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 602.637347] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.637579] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.637729] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.637905] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.638072] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.638288] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.638422] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.638575] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.638734] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.638924] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.639129] env[62952]: DEBUG nova.virt.hardware [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.639898] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe3eb34-2971-4dbc-9158-f2fa3a7e0bf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.648694] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1473c394-d4d2-4a7c-ad4e-04481567d8f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.748202] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.774424] env[62952]: DEBUG nova.compute.manager [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Received event network-changed-70ce59a5-bba1-4503-950e-c3bffab25105 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.774536] env[62952]: DEBUG nova.compute.manager [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Refreshing instance network info cache due to event network-changed-70ce59a5-bba1-4503-950e-c3bffab25105. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 602.774714] env[62952]: DEBUG oslo_concurrency.lockutils [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] Acquiring lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.948519] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Successfully created port: 126d35b6-a8f6-4147-9b91-7ca5da49a4df {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.977771] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367025, 'name': ReconfigVM_Task, 'duration_secs': 0.269123} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.978086] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 602.978774] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-998f6fcd-3319-4bf8-b280-4dae20888b9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.985873] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 602.985873] env[62952]: value = "task-1367026" [ 602.985873] env[62952]: _type = "Task" [ 602.985873] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.994158] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367026, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.161565] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.178051] env[62952]: DEBUG nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 10 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 603.178467] env[62952]: DEBUG nova.compute.provider_tree [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 10 to 11 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 603.179610] env[62952]: DEBUG nova.compute.provider_tree [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 603.238884] env[62952]: DEBUG nova.compute.manager [req-252e9ee8-bceb-44fb-b93f-11ff3eb36b38 req-df2160b1-8fba-4dc5-b2f2-c64be34c8760 service nova] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Received event network-vif-deleted-bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.501536] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367026, 'name': Rename_Task, 'duration_secs': 0.140719} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.502055] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 603.502055] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-705a8502-4240-47e2-a98f-3bdfed154a49 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.508229] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 603.508229] env[62952]: value = "task-1367027" [ 603.508229] env[62952]: _type = "Task" [ 603.508229] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.517819] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367027, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.667056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Releasing lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.667056] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.667056] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.667056] env[62952]: DEBUG oslo_concurrency.lockutils [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] Acquired lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.667056] env[62952]: DEBUG nova.network.neutron [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Refreshing network info cache for port 70ce59a5-bba1-4503-950e-c3bffab25105 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.669756] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de16b97f-6ed7-4248-90f7-01950527afe0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.687100] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6125e79a-7055-4abb-aab7-9da206a1bee2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.697682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.165s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.698197] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.702348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.916s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.702348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.702348] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 603.702348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.366s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.705512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccac552-1029-4b96-9e7b-9ad3ad342b69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.721402] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e98e5c7-e3db-4e10-893a-716fe2769249 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.726229] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e075df95-2102-4f64-b4ba-a65b8fea78d2 could not be found. [ 603.726458] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.726646] env[62952]: INFO nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Took 0.06 seconds to destroy the instance on the hypervisor. [ 603.726888] env[62952]: DEBUG oslo.service.loopingcall [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.727125] env[62952]: DEBUG nova.compute.manager [-] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.727219] env[62952]: DEBUG nova.network.neutron [-] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.739920] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e439bc-560e-4658-adde-5b5685e58f75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.750478] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7adb5de-e2c2-4357-8067-08515e0298d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.786720] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181494MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 603.787031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.796306] env[62952]: DEBUG nova.network.neutron [-] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.020587] env[62952]: DEBUG oslo_vmware.api [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367027, 'name': PowerOnVM_Task, 'duration_secs': 0.422941} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.020587] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 604.020750] env[62952]: INFO nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Took 9.24 seconds to spawn the instance on the hypervisor. [ 604.021225] env[62952]: DEBUG nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 604.022125] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27477700-123c-4d93-bf4e-bfa0297cf111 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.204690] env[62952]: DEBUG nova.compute.utils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.204690] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.206324] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 604.264905] env[62952]: DEBUG nova.network.neutron [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.298767] env[62952]: DEBUG nova.network.neutron [-] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.328638] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Successfully created port: 1f2e3210-fa7e-4658-994a-c243bb1d9cef {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.342509] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9214df-acec-42f1-8ae0-e57b077cd17f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.352061] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a61f12-b1a1-440e-bd0e-d9b14291c71e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.390166] env[62952]: DEBUG nova.policy [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47c2c07c4207491ab2fef673a2810120', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44f05d4de6974d588b5e54c77a2a0731', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.393101] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0999057-6885-41b0-8bf5-41af998722b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.401181] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421888a1-7845-4b21-a7ad-97b6fad4368b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.416282] env[62952]: DEBUG nova.compute.provider_tree [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.554343] env[62952]: INFO nova.compute.manager [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Took 15.65 seconds to build instance. [ 604.598018] env[62952]: DEBUG nova.network.neutron [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.709546] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.802038] env[62952]: INFO nova.compute.manager [-] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Took 1.07 seconds to deallocate network for instance. [ 604.804437] env[62952]: DEBUG nova.compute.claims [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.804616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.924680] env[62952]: DEBUG nova.scheduler.client.report [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.942483] env[62952]: ERROR nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 604.942483] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.942483] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.942483] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.942483] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.942483] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.942483] env[62952]: ERROR nova.compute.manager raise self.value [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.942483] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.942483] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.942483] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.942899] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.942899] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.942899] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 604.942899] env[62952]: ERROR nova.compute.manager [ 604.942899] env[62952]: Traceback (most recent call last): [ 604.942899] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.942899] env[62952]: listener.cb(fileno) [ 604.942899] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.942899] env[62952]: result = function(*args, **kwargs) [ 604.942899] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.942899] env[62952]: return func(*args, **kwargs) [ 604.942899] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.942899] env[62952]: raise e [ 604.942899] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.942899] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 604.942899] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.942899] env[62952]: created_port_ids = self._update_ports_for_instance( [ 604.942899] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.942899] env[62952]: with excutils.save_and_reraise_exception(): [ 604.942899] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.942899] env[62952]: self.force_reraise() [ 604.942899] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.942899] env[62952]: raise self.value [ 604.942899] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.942899] env[62952]: updated_port = self._update_port( [ 604.942899] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.942899] env[62952]: _ensure_no_port_binding_failure(port) [ 604.942899] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.942899] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.943634] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 604.943634] env[62952]: Removing descriptor: 16 [ 604.943634] env[62952]: ERROR nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] Traceback (most recent call last): [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] yield resources [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self.driver.spawn(context, instance, image_meta, [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.943634] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] vm_ref = self.build_virtual_machine(instance, [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] for vif in network_info: [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return self._sync_wrapper(fn, *args, **kwargs) [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self.wait() [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self[:] = self._gt.wait() [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return self._exit_event.wait() [ 604.944151] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] result = hub.switch() [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return self.greenlet.switch() [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] result = function(*args, **kwargs) [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return func(*args, **kwargs) [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] raise e [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] nwinfo = self.network_api.allocate_for_instance( [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.944461] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] created_port_ids = self._update_ports_for_instance( [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] with excutils.save_and_reraise_exception(): [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self.force_reraise() [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] raise self.value [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] updated_port = self._update_port( [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] _ensure_no_port_binding_failure(port) [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.944825] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] raise exception.PortBindingFailed(port_id=port['id']) [ 604.945126] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 604.945126] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] [ 604.945126] env[62952]: INFO nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Terminating instance [ 604.946940] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.947119] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquired lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.947345] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.056633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ae3f7719-ad77-4422-8070-7ae0748ed1a8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "418cd947-00b8-485f-b9fc-2a675c58998d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.164s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.104828] env[62952]: DEBUG oslo_concurrency.lockutils [req-5f749394-b225-4f4b-9b68-4d284c8c29d2 req-d00f2bab-0930-4982-aba4-e9141f70063a service nova] Releasing lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.431796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.727s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.431796] env[62952]: ERROR nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Traceback (most recent call last): [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self.driver.spawn(context, instance, image_meta, [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.431796] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] vm_ref = self.build_virtual_machine(instance, [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] for vif in network_info: [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return self._sync_wrapper(fn, *args, **kwargs) [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self.wait() [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self[:] = self._gt.wait() [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return self._exit_event.wait() [ 605.432096] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] result = hub.switch() [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return self.greenlet.switch() [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] result = function(*args, **kwargs) [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] return func(*args, **kwargs) [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] raise e [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] nwinfo = self.network_api.allocate_for_instance( [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.432514] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] created_port_ids = self._update_ports_for_instance( [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] with excutils.save_and_reraise_exception(): [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] self.force_reraise() [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] raise self.value [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] updated_port = self._update_port( [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] _ensure_no_port_binding_failure(port) [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.432845] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] raise exception.PortBindingFailed(port_id=port['id']) [ 605.433168] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] nova.exception.PortBindingFailed: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. [ 605.433168] env[62952]: ERROR nova.compute.manager [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] [ 605.433168] env[62952]: DEBUG nova.compute.utils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.434425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.647s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.441709] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Build of instance 977675cb-bcf3-431b-aef1-36f8232b093b was re-scheduled: Binding failed for port bbcbf2bf-d35a-49d8-ac1f-07eaca1897a3, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.442220] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.442451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.442600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquired lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.442822] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.503046] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.728726] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.736050] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.770313] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.770605] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.770770] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.770955] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.772307] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.772307] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.772505] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.772642] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.773494] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.774472] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.774562] env[62952]: DEBUG nova.virt.hardware [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.775504] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97af49e-9f12-4901-a0c0-0d35667f363c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.787353] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ed0bea-38fb-469a-8647-d7d5e9903ad5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.965507] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.063914] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Successfully created port: 5f69242e-5593-44bd-b2d1-1b5b4f07153f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.068650] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Successfully created port: f17af0e2-a750-4c58-85f6-1028277f314d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.241657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Releasing lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.241819] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.241930] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.242241] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f230ebfd-04fa-45e8-90a5-1a5fbbacabe0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.245352] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.256378] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09fd4a1-1e0e-4f00-b285-4844654a1f67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.284035] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7f376384-c401-418e-94d3-c9136005a11e could not be found. [ 606.284217] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.284403] env[62952]: INFO nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 606.284649] env[62952]: DEBUG oslo.service.loopingcall [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.284945] env[62952]: DEBUG nova.compute.manager [-] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.285139] env[62952]: DEBUG nova.network.neutron [-] [instance: 7f376384-c401-418e-94d3-c9136005a11e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.317795] env[62952]: DEBUG nova.network.neutron [-] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.748958] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Releasing lock "refresh_cache-977675cb-bcf3-431b-aef1-36f8232b093b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.749262] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.749345] env[62952]: DEBUG nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.749491] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.776188] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.820711] env[62952]: DEBUG nova.network.neutron [-] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.980022] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 977675cb-bcf3-431b-aef1-36f8232b093b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 606.980022] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 418cd947-00b8-485f-b9fc-2a675c58998d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 606.980022] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance e075df95-2102-4f64-b4ba-a65b8fea78d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 606.980022] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 7f376384-c401-418e-94d3-c9136005a11e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 606.980434] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 77418633-2833-47a7-b30f-dca84432d382 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 606.980434] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 21c8106e-8b6c-447b-89e2-df1b34aed12c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 606.980434] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 606.980434] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 607.103274] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f271e1f2-2035-4dd6-95ed-e8985e5bb68c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.112933] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f64488-6f86-4af2-8423-796637bfdb48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.153867] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67df482e-8375-43ef-a6ea-8201398438d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.162294] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5ce7ea-533e-43da-b85d-729cbbcf4602 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.177281] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.279568] env[62952]: DEBUG nova.network.neutron [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.322792] env[62952]: INFO nova.compute.manager [-] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Took 1.04 seconds to deallocate network for instance. [ 607.327959] env[62952]: DEBUG nova.compute.claims [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.328161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.682080] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.783441] env[62952]: INFO nova.compute.manager [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: 977675cb-bcf3-431b-aef1-36f8232b093b] Took 1.03 seconds to deallocate network for instance. [ 608.188125] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 608.190540] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.754s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.190540] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.385s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.515713] env[62952]: DEBUG nova.compute.manager [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Received event network-changed-6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.515937] env[62952]: DEBUG nova.compute.manager [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Refreshing instance network info cache due to event network-changed-6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.516124] env[62952]: DEBUG oslo_concurrency.lockutils [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] Acquiring lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.516259] env[62952]: DEBUG oslo_concurrency.lockutils [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] Acquired lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.516412] env[62952]: DEBUG nova.network.neutron [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Refreshing network info cache for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.834036] env[62952]: INFO nova.scheduler.client.report [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Deleted allocations for instance 977675cb-bcf3-431b-aef1-36f8232b093b [ 608.919145] env[62952]: INFO nova.compute.manager [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Rebuilding instance [ 608.941205] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db06a56-ca3c-42cf-b63a-1ab10f420950 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.951334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f55797f-36fe-4b5a-a5fb-e3ab35b1fcaa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.995645] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308c5301-3771-4501-a172-e0f72b092f46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.005365] env[62952]: DEBUG nova.compute.manager [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 609.006624] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513f48cc-a53f-4239-9984-4976d9462352 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.012180] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ef5d59-73b4-4a17-924e-b28c3f180e84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.025779] env[62952]: DEBUG nova.compute.provider_tree [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.071713] env[62952]: DEBUG nova.network.neutron [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.127197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquiring lock "f569eee5-8806-4493-8625-52b55513f64c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.127427] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Lock "f569eee5-8806-4493-8625-52b55513f64c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.215775] env[62952]: DEBUG nova.network.neutron [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.345993] env[62952]: DEBUG oslo_concurrency.lockutils [None req-edbce28b-435c-4ce0-bdf1-242d77caee7a tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "977675cb-bcf3-431b-aef1-36f8232b093b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.074s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.534212] env[62952]: DEBUG nova.scheduler.client.report [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.547042] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 609.547042] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09490112-d415-47cd-8980-1a6826e06b4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.556618] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 609.556618] env[62952]: value = "task-1367028" [ 609.556618] env[62952]: _type = "Task" [ 609.556618] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.573090] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.630034] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.718997] env[62952]: DEBUG oslo_concurrency.lockutils [req-25081b6f-01d2-4af5-bcd2-2a7fe66a5a3c req-ac15b1c7-9fe9-45b2-983b-ce4ed4dbc0ad service nova] Releasing lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.046844] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.857s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.048916] env[62952]: ERROR nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Traceback (most recent call last): [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self.driver.spawn(context, instance, image_meta, [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] vm_ref = self.build_virtual_machine(instance, [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.048916] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] for vif in network_info: [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return self._sync_wrapper(fn, *args, **kwargs) [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self.wait() [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self[:] = self._gt.wait() [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return self._exit_event.wait() [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] result = hub.switch() [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.049442] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return self.greenlet.switch() [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] result = function(*args, **kwargs) [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] return func(*args, **kwargs) [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] raise e [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] nwinfo = self.network_api.allocate_for_instance( [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] created_port_ids = self._update_ports_for_instance( [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] with excutils.save_and_reraise_exception(): [ 610.049947] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] self.force_reraise() [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] raise self.value [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] updated_port = self._update_port( [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] _ensure_no_port_binding_failure(port) [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] raise exception.PortBindingFailed(port_id=port['id']) [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] nova.exception.PortBindingFailed: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. [ 610.050381] env[62952]: ERROR nova.compute.manager [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] [ 610.051237] env[62952]: DEBUG nova.compute.utils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.052474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.724s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.055297] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Build of instance e075df95-2102-4f64-b4ba-a65b8fea78d2 was re-scheduled: Binding failed for port 70ce59a5-bba1-4503-950e-c3bffab25105, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.055760] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.055983] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquiring lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.056156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Acquired lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.056337] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.078844] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367028, 'name': PowerOffVM_Task, 'duration_secs': 0.118983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.078844] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 610.078844] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.078844] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc22acb-3e4d-4ebe-a7af-f71573cb4f9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.086387] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 610.086676] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d225d381-698f-46c9-954d-e65bf0f05575 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.114083] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 610.114303] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 610.114482] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleting the datastore file [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 610.114731] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c08a5e05-53f1-4c94-a4aa-95c53c8c5f4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.124512] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 610.124512] env[62952]: value = "task-1367030" [ 610.124512] env[62952]: _type = "Task" [ 610.124512] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.134960] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.163216] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.433425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquiring lock "f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.433621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Lock "f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.638823] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104035} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.641991] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 610.642945] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 610.643158] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.659318] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.722815] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe1b914-b54a-4cb3-8700-0b2ab1869343 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.735359] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeac9043-ce2b-4e08-8d1e-904c097ae182 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.773204] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c85919-65fd-4b13-83ab-d1228b7d9cbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.781302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c1a733-34cb-4e06-9aac-bb5b1f69af91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.797119] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 610.824585] env[62952]: ERROR nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 610.824585] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.824585] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.824585] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.824585] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.824585] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.824585] env[62952]: ERROR nova.compute.manager raise self.value [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.824585] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.824585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.824585] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.825052] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.825052] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.825052] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 610.825052] env[62952]: ERROR nova.compute.manager [ 610.825052] env[62952]: Traceback (most recent call last): [ 610.825052] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.825052] env[62952]: listener.cb(fileno) [ 610.825052] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.825052] env[62952]: result = function(*args, **kwargs) [ 610.825052] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.825052] env[62952]: return func(*args, **kwargs) [ 610.825052] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.825052] env[62952]: raise e [ 610.825052] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.825052] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 610.825052] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.825052] env[62952]: created_port_ids = self._update_ports_for_instance( [ 610.825052] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.825052] env[62952]: with excutils.save_and_reraise_exception(): [ 610.825052] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.825052] env[62952]: self.force_reraise() [ 610.825052] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.825052] env[62952]: raise self.value [ 610.825052] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.825052] env[62952]: updated_port = self._update_port( [ 610.825052] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.825052] env[62952]: _ensure_no_port_binding_failure(port) [ 610.825052] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.825052] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.825739] env[62952]: nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 610.825739] env[62952]: Removing descriptor: 19 [ 610.825739] env[62952]: ERROR nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Traceback (most recent call last): [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] yield resources [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self.driver.spawn(context, instance, image_meta, [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.825739] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] vm_ref = self.build_virtual_machine(instance, [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] for vif in network_info: [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return self._sync_wrapper(fn, *args, **kwargs) [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self.wait() [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self[:] = self._gt.wait() [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return self._exit_event.wait() [ 610.826051] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] result = hub.switch() [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return self.greenlet.switch() [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] result = function(*args, **kwargs) [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return func(*args, **kwargs) [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] raise e [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] nwinfo = self.network_api.allocate_for_instance( [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.826373] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] created_port_ids = self._update_ports_for_instance( [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] with excutils.save_and_reraise_exception(): [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self.force_reraise() [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] raise self.value [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] updated_port = self._update_port( [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] _ensure_no_port_binding_failure(port) [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.826695] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] raise exception.PortBindingFailed(port_id=port['id']) [ 610.826999] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 610.826999] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] [ 610.826999] env[62952]: INFO nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Terminating instance [ 610.831255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquiring lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.831255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquired lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.831255] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.935854] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.962567] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.980194] env[62952]: ERROR nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 610.980194] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.980194] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.980194] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.980194] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.980194] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.980194] env[62952]: ERROR nova.compute.manager raise self.value [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.980194] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.980194] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.980194] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.980643] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.980643] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.980643] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 610.980643] env[62952]: ERROR nova.compute.manager [ 610.980643] env[62952]: Traceback (most recent call last): [ 610.980643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.980643] env[62952]: listener.cb(fileno) [ 610.980643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.980643] env[62952]: result = function(*args, **kwargs) [ 610.980643] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.980643] env[62952]: return func(*args, **kwargs) [ 610.980643] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.980643] env[62952]: raise e [ 610.980643] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.980643] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 610.980643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.980643] env[62952]: created_port_ids = self._update_ports_for_instance( [ 610.980643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.980643] env[62952]: with excutils.save_and_reraise_exception(): [ 610.980643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.980643] env[62952]: self.force_reraise() [ 610.980643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.980643] env[62952]: raise self.value [ 610.980643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.980643] env[62952]: updated_port = self._update_port( [ 610.980643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.980643] env[62952]: _ensure_no_port_binding_failure(port) [ 610.980643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.980643] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.981409] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 610.981409] env[62952]: Removing descriptor: 15 [ 610.981409] env[62952]: ERROR nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] Traceback (most recent call last): [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] yield resources [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self.driver.spawn(context, instance, image_meta, [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.981409] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] vm_ref = self.build_virtual_machine(instance, [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] for vif in network_info: [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return self._sync_wrapper(fn, *args, **kwargs) [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self.wait() [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self[:] = self._gt.wait() [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return self._exit_event.wait() [ 610.981713] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] result = hub.switch() [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return self.greenlet.switch() [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] result = function(*args, **kwargs) [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return func(*args, **kwargs) [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] raise e [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] nwinfo = self.network_api.allocate_for_instance( [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.982157] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] created_port_ids = self._update_ports_for_instance( [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] with excutils.save_and_reraise_exception(): [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self.force_reraise() [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] raise self.value [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] updated_port = self._update_port( [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] _ensure_no_port_binding_failure(port) [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.982491] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] raise exception.PortBindingFailed(port_id=port['id']) [ 610.982793] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 610.982793] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] [ 610.982793] env[62952]: INFO nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Terminating instance [ 610.984494] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.984494] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquired lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.984494] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.332530] env[62952]: ERROR nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [req-2943e3b2-cc46-43b6-9202-74abab14586d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2943e3b2-cc46-43b6-9202-74abab14586d"}]}: nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 611.359787] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 611.377048] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.379552] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 611.379760] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 611.393926] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 611.417648] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 611.466627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.467403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Releasing lock "refresh_cache-e075df95-2102-4f64-b4ba-a65b8fea78d2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.467403] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.467490] env[62952]: DEBUG nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.467635] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.493182] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.571077] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ed82da-049e-4db6-bd9c-f526ea995a73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.581846] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf48ac3-af68-4eba-b9a4-aecac5e1e279 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.618969] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.624086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a86b74-f5af-455c-8eb3-46be199ae701 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.630991] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e414a7-5394-43f8-af75-8e64a1b19449 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.646756] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 611.709150] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.709456] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.709632] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.709845] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.710058] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.710158] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.710387] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.710537] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.710759] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.710840] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.711060] env[62952]: DEBUG nova.virt.hardware [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.714584] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb53b184-7a13-4a75-a00c-355e9606a035 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.725165] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.729723] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5777eca8-e13e-43a5-a2f6-c4629e533ad4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.749570] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 611.761016] env[62952]: DEBUG oslo.service.loopingcall [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.761016] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 611.764137] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9437f5ec-634b-4c9c-a2b9-e07ea4f84112 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.781178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquiring lock "51db1590-9612-4d23-a49f-e1900f82ed90" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.781416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Lock "51db1590-9612-4d23-a49f-e1900f82ed90" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.794400] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 611.794400] env[62952]: value = "task-1367031" [ 611.794400] env[62952]: _type = "Task" [ 611.794400] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.801565] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367031, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.865424] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.996200] env[62952]: DEBUG nova.network.neutron [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.179250] env[62952]: ERROR nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [req-2a0c2726-5b75-4e4c-82ec-cf3c3dce594e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2a0c2726-5b75-4e4c-82ec-cf3c3dce594e"}]}: nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 612.202416] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 612.219120] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 612.219120] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 612.230968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Releasing lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.230968] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.230968] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.230968] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef6b8ab4-31ca-47bd-9a85-510d6113821f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.239477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d71b51-2940-4fab-ad30-1dc844444c7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.254121] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 612.276194] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21c8106e-8b6c-447b-89e2-df1b34aed12c could not be found. [ 612.276414] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.276596] env[62952]: INFO nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 612.276831] env[62952]: DEBUG oslo.service.loopingcall [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.277086] env[62952]: DEBUG nova.compute.manager [-] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.277188] env[62952]: DEBUG nova.network.neutron [-] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.283184] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 612.286578] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.308791] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367031, 'name': CreateVM_Task, 'duration_secs': 0.396693} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.308791] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 612.310338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.310338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.310338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.310606] env[62952]: DEBUG nova.network.neutron [-] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.312099] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3466abd2-f625-4ad5-8d45-2ed8f0c34e0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.316921] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 612.316921] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52fa9d22-5395-e456-d31e-5f48699a5c60" [ 612.316921] env[62952]: _type = "Task" [ 612.316921] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.328808] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fa9d22-5395-e456-d31e-5f48699a5c60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.375080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Releasing lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.375624] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.375750] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.376759] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5265544-a203-41ec-88b1-760d96f7d72f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.396529] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a029e9b4-35b0-4df5-ab7c-ae1ac25bd09b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.428307] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 77418633-2833-47a7-b30f-dca84432d382 could not be found. [ 612.428307] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.428307] env[62952]: INFO nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Took 0.05 seconds to destroy the instance on the hypervisor. [ 612.428307] env[62952]: DEBUG oslo.service.loopingcall [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.428307] env[62952]: DEBUG nova.compute.manager [-] [instance: 77418633-2833-47a7-b30f-dca84432d382] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.428307] env[62952]: DEBUG nova.network.neutron [-] [instance: 77418633-2833-47a7-b30f-dca84432d382] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.441584] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquiring lock "1dc0311e-2f03-402a-bb31-117289502ab0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.441795] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Lock "1dc0311e-2f03-402a-bb31-117289502ab0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.452203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquiring lock "6511f90d-82e8-4374-bdaf-0a6cc66a5a65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.452492] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Lock "6511f90d-82e8-4374-bdaf-0a6cc66a5a65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.500407] env[62952]: INFO nova.compute.manager [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] [instance: e075df95-2102-4f64-b4ba-a65b8fea78d2] Took 1.03 seconds to deallocate network for instance. [ 612.547303] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618d14f4-204a-42d2-b9a1-e8951c81c3cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.557221] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb50f9c7-9d1a-4042-86ed-d9049731e853 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.592786] env[62952]: DEBUG nova.network.neutron [-] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.595290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0acace-fb34-446e-9fd7-2a07a3681e53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.603239] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728d2382-09f3-48fc-b947-397e71bfddeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.618245] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 612.814870] env[62952]: DEBUG nova.network.neutron [-] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.832641] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fa9d22-5395-e456-d31e-5f48699a5c60, 'name': SearchDatastore_Task, 'duration_secs': 0.011797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.832862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.834416] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 612.834416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.834416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.834416] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 612.834416] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7d3c0a3-b7d1-4c31-b479-98aeabcd72d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.844016] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.844265] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 612.845199] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09f41329-fcf5-47df-98a4-07800c616e5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.849392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.857195] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 612.857195] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b7bdfa-5896-0c54-339e-91710514cf3f" [ 612.857195] env[62952]: _type = "Task" [ 612.857195] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.866641] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b7bdfa-5896-0c54-339e-91710514cf3f, 'name': SearchDatastore_Task, 'duration_secs': 0.008648} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.867551] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01e72762-be00-4618-87f6-b79716d83863 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.873636] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 612.873636] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ac99af-f5cc-d921-9a8d-d5a862459b2c" [ 612.873636] env[62952]: _type = "Task" [ 612.873636] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.883780] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ac99af-f5cc-d921-9a8d-d5a862459b2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.911787] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquiring lock "722f0334-6210-46b8-87de-d9bf24812a1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.912086] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Lock "722f0334-6210-46b8-87de-d9bf24812a1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.947768] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.956711] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.048451] env[62952]: DEBUG nova.compute.manager [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Received event network-changed-f17af0e2-a750-4c58-85f6-1028277f314d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.048656] env[62952]: DEBUG nova.compute.manager [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Refreshing instance network info cache due to event network-changed-f17af0e2-a750-4c58-85f6-1028277f314d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.048865] env[62952]: DEBUG oslo_concurrency.lockutils [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] Acquiring lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.049011] env[62952]: DEBUG oslo_concurrency.lockutils [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] Acquired lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.049254] env[62952]: DEBUG nova.network.neutron [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Refreshing network info cache for port f17af0e2-a750-4c58-85f6-1028277f314d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 613.161621] env[62952]: ERROR nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [req-10f5e102-ec2e-4e9f-bf9e-a54db8b2f152] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-10f5e102-ec2e-4e9f-bf9e-a54db8b2f152"}]}: nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 613.202402] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 613.222107] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 613.222107] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 613.237766] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 613.265965] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 613.318938] env[62952]: INFO nova.compute.manager [-] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Took 1.04 seconds to deallocate network for instance. [ 613.320941] env[62952]: DEBUG nova.compute.claims [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.321132] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.395340] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ac99af-f5cc-d921-9a8d-d5a862459b2c, 'name': SearchDatastore_Task, 'duration_secs': 0.008393} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.397832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.397919] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 613.401419] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc61237d-974f-4433-b5f2-c14dced67b7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.406275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "510731f2-399d-486a-8c60-1c61c05063de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.406275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "510731f2-399d-486a-8c60-1c61c05063de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.411840] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 613.411840] env[62952]: value = "task-1367032" [ 613.411840] env[62952]: _type = "Task" [ 613.411840] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.421072] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.434829] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.491087] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.492451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.499072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquiring lock "7c98a241-9297-417a-b412-5c279144572b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.499072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Lock "7c98a241-9297-417a-b412-5c279144572b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.546585] env[62952]: INFO nova.scheduler.client.report [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Deleted allocations for instance e075df95-2102-4f64-b4ba-a65b8fea78d2 [ 613.560836] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a1a220-52e7-4b3b-ab78-102acd076850 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.571414] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b907ca3c-e5e6-43b1-bb1a-dc6a1041f8c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.620633] env[62952]: DEBUG nova.network.neutron [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.622873] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2647686c-6ea3-4cd0-ac1e-ebc891417869 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.635576] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2a7a69-f720-4bc8-80b0-415e9471492e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.657853] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 613.821356] env[62952]: DEBUG nova.network.neutron [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.926065] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367032, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.937382] env[62952]: DEBUG nova.network.neutron [-] [instance: 77418633-2833-47a7-b30f-dca84432d382] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.953810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.065547] env[62952]: DEBUG oslo_concurrency.lockutils [None req-46d698f2-9899-4025-bdfd-30f453da70ae tempest-FloatingIPsAssociationTestJSON-108032633 tempest-FloatingIPsAssociationTestJSON-108032633-project-member] Lock "e075df95-2102-4f64-b4ba-a65b8fea78d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.695s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.214436] env[62952]: DEBUG nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 19 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 614.215561] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 19 to 20 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 614.220313] env[62952]: DEBUG nova.compute.provider_tree [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 614.325111] env[62952]: DEBUG oslo_concurrency.lockutils [req-7f41cc6a-aaee-4224-9273-68647954a0ef req-2392c8fd-fa3c-4f85-a066-2f1b24904274 service nova] Releasing lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.431451] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367032, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52634} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.431451] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.431451] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.431451] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b4c21f6-e5c1-4e14-85c4-1509c739a217 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.435822] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 614.435822] env[62952]: value = "task-1367033" [ 614.435822] env[62952]: _type = "Task" [ 614.435822] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.442313] env[62952]: INFO nova.compute.manager [-] [instance: 77418633-2833-47a7-b30f-dca84432d382] Took 2.01 seconds to deallocate network for instance. [ 614.445423] env[62952]: DEBUG nova.compute.claims [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.445423] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.448709] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.571072] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.725884] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 4.673s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.726133] env[62952]: ERROR nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] Traceback (most recent call last): [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self.driver.spawn(context, instance, image_meta, [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] vm_ref = self.build_virtual_machine(instance, [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.726133] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] for vif in network_info: [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return self._sync_wrapper(fn, *args, **kwargs) [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self.wait() [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self[:] = self._gt.wait() [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return self._exit_event.wait() [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] result = hub.switch() [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.726503] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return self.greenlet.switch() [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] result = function(*args, **kwargs) [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] return func(*args, **kwargs) [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] raise e [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] nwinfo = self.network_api.allocate_for_instance( [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] created_port_ids = self._update_ports_for_instance( [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] with excutils.save_and_reraise_exception(): [ 614.726882] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] self.force_reraise() [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] raise self.value [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] updated_port = self._update_port( [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] _ensure_no_port_binding_failure(port) [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] raise exception.PortBindingFailed(port_id=port['id']) [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] nova.exception.PortBindingFailed: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. [ 614.727315] env[62952]: ERROR nova.compute.manager [instance: 7f376384-c401-418e-94d3-c9136005a11e] [ 614.729267] env[62952]: DEBUG nova.compute.utils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.731256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.568s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.733517] env[62952]: INFO nova.compute.claims [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.737709] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Build of instance 7f376384-c401-418e-94d3-c9136005a11e was re-scheduled: Binding failed for port 6b5a1ca5-3cb2-47c1-89ea-96b6fe567bb9, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.738641] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.739707] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.740197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquired lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.742954] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.945911] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074281} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.946084] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.946842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de974b8-2f56-49d5-98f2-4c2ef6f63f24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.968714] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.968714] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6fb51c3-b107-45fa-8d3e-0122dd155cfd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.991015] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 614.991015] env[62952]: value = "task-1367034" [ 614.991015] env[62952]: _type = "Task" [ 614.991015] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.000939] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367034, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.102582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.193322] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquiring lock "542169bf-71e2-47f8-a311-8a3dca4ffee7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.193527] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Lock "542169bf-71e2-47f8-a311-8a3dca4ffee7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.232905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "12408f00-ed1d-4b96-bc0c-59779d0eda74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.233143] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "12408f00-ed1d-4b96-bc0c-59779d0eda74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.268905] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.330034] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.502393] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367034, 'name': ReconfigVM_Task, 'duration_secs': 0.488719} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.502698] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 615.503492] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd5ffe42-0f63-43b3-a1a5-f26921c84e6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.513273] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 615.513273] env[62952]: value = "task-1367035" [ 615.513273] env[62952]: _type = "Task" [ 615.513273] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.525457] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367035, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.833651] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Releasing lock "refresh_cache-7f376384-c401-418e-94d3-c9136005a11e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.833651] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.833651] env[62952]: DEBUG nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.833781] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.866080] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.999973] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4beb29c7-ffd7-45c0-bc40-fc702b63cb1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.008070] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390c6766-9d88-4a13-bc22-4cb32f6eb2ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.045256] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1d50b2-a1e1-4213-8e0c-d1b3dfaac791 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.048858] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367035, 'name': Rename_Task, 'duration_secs': 0.131212} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.049537] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.049922] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fcb8630-deab-4059-ba4e-9d16f98ab79a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.055551] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05cb5c6-1833-4aca-889f-9f6b712b7445 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.060534] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 616.060534] env[62952]: value = "task-1367036" [ 616.060534] env[62952]: _type = "Task" [ 616.060534] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.072533] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 616.076377] env[62952]: DEBUG nova.compute.manager [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 77418633-2833-47a7-b30f-dca84432d382] Received event network-changed-126d35b6-a8f6-4147-9b91-7ca5da49a4df {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.076576] env[62952]: DEBUG nova.compute.manager [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 77418633-2833-47a7-b30f-dca84432d382] Refreshing instance network info cache due to event network-changed-126d35b6-a8f6-4147-9b91-7ca5da49a4df. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 616.076869] env[62952]: DEBUG oslo_concurrency.lockutils [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] Acquiring lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.076952] env[62952]: DEBUG oslo_concurrency.lockutils [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] Acquired lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.077140] env[62952]: DEBUG nova.network.neutron [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 77418633-2833-47a7-b30f-dca84432d382] Refreshing network info cache for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 616.082566] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367036, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.371800] env[62952]: DEBUG nova.network.neutron [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.573950] env[62952]: DEBUG oslo_vmware.api [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367036, 'name': PowerOnVM_Task, 'duration_secs': 0.4623} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.574658] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.575077] env[62952]: DEBUG nova.compute.manager [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 616.576214] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bce42d8-32b7-4d5f-90d2-95d46d92f753 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.611968] env[62952]: ERROR nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [req-c430b425-0fd7-416c-9715-228c16e72716] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c430b425-0fd7-416c-9715-228c16e72716"}]} [ 616.632086] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 616.635636] env[62952]: DEBUG nova.network.neutron [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.651989] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 616.652345] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 616.672104] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 616.692377] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 616.873681] env[62952]: DEBUG nova.network.neutron [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 77418633-2833-47a7-b30f-dca84432d382] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.874449] env[62952]: INFO nova.compute.manager [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 7f376384-c401-418e-94d3-c9136005a11e] Took 1.04 seconds to deallocate network for instance. [ 616.926370] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5433c59-1643-4281-b16b-eadf89f3762b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.936913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7ec6d0-9a52-4f1f-9e5f-4f8583d2f583 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.971704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966c9f14-8a5c-4a44-a672-afa6e6872500 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.981024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6e73ec-f0af-40d6-ba42-be9ddc3c7631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.996269] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 617.108177] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.390100] env[62952]: DEBUG oslo_concurrency.lockutils [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] Releasing lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.390100] env[62952]: DEBUG nova.compute.manager [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Received event network-vif-deleted-f17af0e2-a750-4c58-85f6-1028277f314d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.390100] env[62952]: DEBUG nova.compute.manager [req-22978a13-d285-4bdf-9b94-40d9b6bd247a req-6369fbf3-ae1a-4ad1-acf1-2b63f798b0b4 service nova] [instance: 77418633-2833-47a7-b30f-dca84432d382] Received event network-vif-deleted-126d35b6-a8f6-4147-9b91-7ca5da49a4df {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.531285] env[62952]: ERROR nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [req-39602d1b-8da7-4aa4-bbb7-a47e02f8e3ef] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-39602d1b-8da7-4aa4-bbb7-a47e02f8e3ef"}]} [ 617.557279] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 617.577027] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 617.577428] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 617.594808] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 617.617995] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 617.833305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "288918a3-85be-416d-99fc-fd3450029749" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.833543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "288918a3-85be-416d-99fc-fd3450029749" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.856027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b60832-91b4-45ea-b765-73062209b884 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.863535] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "d8653006-b429-44b5-b978-ee00a279e50e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.863862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "d8653006-b429-44b5-b978-ee00a279e50e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.868694] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278034db-82e5-4d39-ac39-f4d777a60d95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.907268] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90d3249-7a0a-4595-b707-9bb836c2e984 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.915366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8c31db-669a-4b20-a1ca-9ec2f13f6bd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.931045] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 617.939954] env[62952]: INFO nova.scheduler.client.report [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Deleted allocations for instance 7f376384-c401-418e-94d3-c9136005a11e [ 618.450101] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6e8c2306-5d78-4aff-a914-274b7effca92 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "7f376384-c401-418e-94d3-c9136005a11e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.185s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.483335] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 24 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 618.483584] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 24 to 25 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 618.483766] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 618.951783] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.990231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.259s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.990797] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 618.994114] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.528s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.996249] env[62952]: INFO nova.compute.claims [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.483681] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.502052] env[62952]: DEBUG nova.compute.utils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.504611] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.504611] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.623867] env[62952]: DEBUG nova.policy [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7663205fe1db4c0282f62b97468813ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db873ea7843e45a1be7f9eae63c0a10f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.004924] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.293024] env[62952]: INFO nova.compute.manager [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Rebuilding instance [ 620.325592] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54930aa4-1e45-41dd-bdc1-ba41add4da07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.341922] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb955fa6-adb8-423f-a365-be9a4de2cf64 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.382242] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbba49c5-711d-441c-a727-d49dd5cf3aff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.394779] env[62952]: DEBUG nova.compute.manager [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 620.395630] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c374d1-c5c6-4140-a321-1774ac09d8d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.403195] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66f1873-8669-4b82-ac7b-3eb2613298a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.424265] env[62952]: DEBUG nova.compute.provider_tree [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.874256] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Successfully created port: 288e9672-8694-49b1-aea7-8532f1373dec {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.914563] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 620.914563] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95dfd20f-3686-4933-8fc2-d7ff40c92ced {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.922808] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 620.922808] env[62952]: value = "task-1367041" [ 620.922808] env[62952]: _type = "Task" [ 620.922808] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.926840] env[62952]: DEBUG nova.scheduler.client.report [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.936587] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.026092] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.056044] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.056044] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.056232] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.058163] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.058163] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.058163] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.058163] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.058163] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.058400] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.058400] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.058400] env[62952]: DEBUG nova.virt.hardware [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.058495] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae52483-fe3d-4f43-8e37-7b0aa5774673 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.067216] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6774e4-025f-4bcb-8bfa-8341247d2b92 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.432208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.432754] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.439470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.590s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.441464] env[62952]: INFO nova.compute.claims [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.443541] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367041, 'name': PowerOffVM_Task, 'duration_secs': 0.119724} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.444119] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 621.448019] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 621.448019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f23ac1-f504-4fd8-a7f4-7176492635a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.453209] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 621.453450] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66577111-5318-4086-a7ac-cfce4de1e4ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.481955] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 621.481955] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 621.481955] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Deleting the datastore file [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 621.481955] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87fd2008-fc75-4917-bf4b-ea367e3b4eec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.491365] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 621.491365] env[62952]: value = "task-1367043" [ 621.491365] env[62952]: _type = "Task" [ 621.491365] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.503046] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367043, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.948605] env[62952]: DEBUG nova.compute.utils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.951597] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.951778] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.007782] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367043, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093031} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.007782] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 622.007782] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 622.007782] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.140280] env[62952]: DEBUG nova.policy [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0234959d936347328283732709d5dc8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b297c51d29e4419eaed16d4642f04f40', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.457619] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.624119] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquiring lock "8db47d06-230f-4c5e-8bd6-9d5b543455b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.624333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Lock "8db47d06-230f-4c5e-8bd6-9d5b543455b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.830730] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d7c9bb-d78e-4637-9bf6-fe42aa99ef0f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.839841] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46281406-132d-4ed5-b402-24b64d2ad4c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.882970] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5a76c9-3077-4d89-902c-f1ab3b9ad1be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.890698] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8784bc5-e860-4557-a418-227becdae5c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.904551] env[62952]: DEBUG nova.compute.provider_tree [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 623.056157] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.056304] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.056740] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.056740] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.056740] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.056899] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.057131] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.057316] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.057784] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.057784] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.057890] env[62952]: DEBUG nova.virt.hardware [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.059184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251334e2-8e38-4471-b13f-f7e794e3cf29 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.073788] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f6e7fc-9014-4ab5-af96-797d83ebc4a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.093732] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 623.102464] env[62952]: DEBUG oslo.service.loopingcall [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.103068] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 623.103068] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-457a6a96-8049-432a-9f78-3d08dbacfa4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.122269] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 623.122269] env[62952]: value = "task-1367045" [ 623.122269] env[62952]: _type = "Task" [ 623.122269] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.132967] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367045, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.187074] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquiring lock "f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.187074] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Lock "f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.289781] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Successfully created port: 201ac188-f487-4348-97ee-c7533f56677f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.430503] env[62952]: ERROR nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [req-b15c07e9-1929-4e8a-94ec-308f3651b66e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b15c07e9-1929-4e8a-94ec-308f3651b66e"}]} [ 623.461696] env[62952]: DEBUG nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 623.475361] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.481293] env[62952]: DEBUG nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 623.481832] env[62952]: DEBUG nova.compute.provider_tree [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 623.509054] env[62952]: DEBUG nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 623.516280] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.516280] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.516454] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.516505] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.516646] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.516787] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.516985] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.517143] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.517301] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.517455] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.517663] env[62952]: DEBUG nova.virt.hardware [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.522031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc83d89-007d-4f16-bc82-8cabe94cbdab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.533895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f9dc61-5c53-49d4-9612-f07788dd07d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.539024] env[62952]: DEBUG nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 623.642135] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367045, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.930148] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e380c82e-64b5-481d-9446-f159197be5be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.941018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9ba1c7-7cdd-4da7-b78e-c18fc7544727 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.978218] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545b63ac-24a0-4f0e-931d-425e6f72d6de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.996762] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bd84a8-d2ca-4fd3-8b1a-37f7ee62abfd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.013371] env[62952]: DEBUG nova.compute.provider_tree [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 624.143571] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367045, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.568356] env[62952]: DEBUG nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 27 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 624.569192] env[62952]: DEBUG nova.compute.provider_tree [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 27 to 28 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 624.569192] env[62952]: DEBUG nova.compute.provider_tree [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 624.643717] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367045, 'name': CreateVM_Task, 'duration_secs': 1.299298} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.644266] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 624.644806] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.645090] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.645363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 624.645611] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfd8f066-8b19-4055-99e2-c28e5c130ba6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.651968] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 624.651968] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5236cdd7-f5d7-7db4-8094-af56b8d72a32" [ 624.651968] env[62952]: _type = "Task" [ 624.651968] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.661244] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5236cdd7-f5d7-7db4-8094-af56b8d72a32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.076946] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.637s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.077574] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.084168] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.761s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.097334] env[62952]: ERROR nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 625.097334] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.097334] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.097334] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.097334] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.097334] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.097334] env[62952]: ERROR nova.compute.manager raise self.value [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.097334] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.097334] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.097334] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.097954] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.097954] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.097954] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 625.097954] env[62952]: ERROR nova.compute.manager [ 625.097954] env[62952]: Traceback (most recent call last): [ 625.097954] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.097954] env[62952]: listener.cb(fileno) [ 625.097954] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.097954] env[62952]: result = function(*args, **kwargs) [ 625.097954] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.097954] env[62952]: return func(*args, **kwargs) [ 625.097954] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.097954] env[62952]: raise e [ 625.097954] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.097954] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 625.097954] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.097954] env[62952]: created_port_ids = self._update_ports_for_instance( [ 625.097954] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.097954] env[62952]: with excutils.save_and_reraise_exception(): [ 625.097954] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.097954] env[62952]: self.force_reraise() [ 625.097954] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.097954] env[62952]: raise self.value [ 625.097954] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.097954] env[62952]: updated_port = self._update_port( [ 625.097954] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.097954] env[62952]: _ensure_no_port_binding_failure(port) [ 625.097954] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.097954] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.099734] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 625.099734] env[62952]: Removing descriptor: 19 [ 625.099734] env[62952]: ERROR nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] Traceback (most recent call last): [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] yield resources [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self.driver.spawn(context, instance, image_meta, [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.099734] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] vm_ref = self.build_virtual_machine(instance, [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] for vif in network_info: [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return self._sync_wrapper(fn, *args, **kwargs) [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self.wait() [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self[:] = self._gt.wait() [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return self._exit_event.wait() [ 625.100733] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] result = hub.switch() [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return self.greenlet.switch() [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] result = function(*args, **kwargs) [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return func(*args, **kwargs) [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] raise e [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] nwinfo = self.network_api.allocate_for_instance( [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.102114] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] created_port_ids = self._update_ports_for_instance( [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] with excutils.save_and_reraise_exception(): [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self.force_reraise() [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] raise self.value [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] updated_port = self._update_port( [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] _ensure_no_port_binding_failure(port) [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.102517] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] raise exception.PortBindingFailed(port_id=port['id']) [ 625.102871] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 625.102871] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] [ 625.102871] env[62952]: INFO nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Terminating instance [ 625.105899] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquiring lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.105899] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquired lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.105899] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.162256] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5236cdd7-f5d7-7db4-8094-af56b8d72a32, 'name': SearchDatastore_Task, 'duration_secs': 0.009147} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.162541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.162766] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.162993] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.163324] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.163399] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.163874] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63b79f0f-004d-4fa4-9a15-fb6ef77cd921 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.171702] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.171874] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.172570] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86288acd-2049-467e-98de-59e035aba4e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.177901] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 625.177901] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5296a31b-a6d8-a543-3b8a-4cd47135503a" [ 625.177901] env[62952]: _type = "Task" [ 625.177901] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.186289] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5296a31b-a6d8-a543-3b8a-4cd47135503a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.557852] env[62952]: DEBUG nova.compute.manager [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] [instance: f569eee5-8806-4493-8625-52b55513f64c] Received event network-changed-288e9672-8694-49b1-aea7-8532f1373dec {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.558569] env[62952]: DEBUG nova.compute.manager [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] [instance: f569eee5-8806-4493-8625-52b55513f64c] Refreshing instance network info cache due to event network-changed-288e9672-8694-49b1-aea7-8532f1373dec. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.558992] env[62952]: DEBUG oslo_concurrency.lockutils [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] Acquiring lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.585994] env[62952]: DEBUG nova.compute.utils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.585994] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.590981] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.653610] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.700741] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5296a31b-a6d8-a543-3b8a-4cd47135503a, 'name': SearchDatastore_Task, 'duration_secs': 0.007505} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.702110] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98fea8cb-4772-4c86-9bc0-575a88b37945 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.708322] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 625.708322] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ad0c0a-dab2-70e4-2f08-92ca05f1b23a" [ 625.708322] env[62952]: _type = "Task" [ 625.708322] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.721978] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ad0c0a-dab2-70e4-2f08-92ca05f1b23a, 'name': SearchDatastore_Task, 'duration_secs': 0.00939} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.722292] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.722787] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.722947] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e31194e-72b3-4136-a978-9fa33111fbf7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.735008] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 625.735008] env[62952]: value = "task-1367047" [ 625.735008] env[62952]: _type = "Task" [ 625.735008] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.741777] env[62952]: DEBUG nova.policy [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ee426170c494abfb8d228035a1da688', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '451b42f86347460fa0c30e51334de0c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.751840] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.994939] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.045231] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aeda8b8-3eb1-4481-a532-8415663c85b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.062403] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9b6788-4131-447c-bb09-6c91d26470a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.107064] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.110131] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0946d965-42ff-4f67-98ea-8efd6c0ab2c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.121825] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb89d9e-c419-47ce-b1d9-9ac8e9e9341a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.139368] env[62952]: DEBUG nova.compute.provider_tree [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.224334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "40b6590d-44e1-4d02-b1fa-46c1e6c861ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.224570] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "40b6590d-44e1-4d02-b1fa-46c1e6c861ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.249690] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483754} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.249994] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 626.250229] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 626.250489] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4126a5f-4cd6-4097-9f40-89be29b76294 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.259615] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 626.259615] env[62952]: value = "task-1367048" [ 626.259615] env[62952]: _type = "Task" [ 626.259615] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.269651] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.504155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Releasing lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.504676] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.504971] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.505399] env[62952]: DEBUG oslo_concurrency.lockutils [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] Acquired lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.505582] env[62952]: DEBUG nova.network.neutron [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] [instance: f569eee5-8806-4493-8625-52b55513f64c] Refreshing network info cache for port 288e9672-8694-49b1-aea7-8532f1373dec {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.508047] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbfcd1cf-c018-4bb2-b321-d26dfc277c10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.516709] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d4439f-9588-44d8-b1a6-b0f72c68710f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.542080] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f569eee5-8806-4493-8625-52b55513f64c could not be found. [ 626.542080] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.542080] env[62952]: INFO nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.542387] env[62952]: DEBUG oslo.service.loopingcall [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.542619] env[62952]: DEBUG nova.compute.manager [-] [instance: f569eee5-8806-4493-8625-52b55513f64c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.542619] env[62952]: DEBUG nova.network.neutron [-] [instance: f569eee5-8806-4493-8625-52b55513f64c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.587637] env[62952]: DEBUG nova.network.neutron [-] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.642806] env[62952]: DEBUG nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.680646] env[62952]: ERROR nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 626.680646] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.680646] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.680646] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.680646] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.680646] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.680646] env[62952]: ERROR nova.compute.manager raise self.value [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.680646] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.680646] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.680646] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.681371] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.681371] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.681371] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 626.681371] env[62952]: ERROR nova.compute.manager [ 626.681371] env[62952]: Traceback (most recent call last): [ 626.681371] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.681371] env[62952]: listener.cb(fileno) [ 626.681371] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.681371] env[62952]: result = function(*args, **kwargs) [ 626.681371] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.681371] env[62952]: return func(*args, **kwargs) [ 626.681371] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.681371] env[62952]: raise e [ 626.681371] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.681371] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 626.681371] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.681371] env[62952]: created_port_ids = self._update_ports_for_instance( [ 626.681371] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.681371] env[62952]: with excutils.save_and_reraise_exception(): [ 626.681371] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.681371] env[62952]: self.force_reraise() [ 626.681371] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.681371] env[62952]: raise self.value [ 626.681371] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.681371] env[62952]: updated_port = self._update_port( [ 626.681371] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.681371] env[62952]: _ensure_no_port_binding_failure(port) [ 626.681371] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.681371] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.683725] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 626.683725] env[62952]: Removing descriptor: 15 [ 626.683725] env[62952]: ERROR nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Traceback (most recent call last): [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] yield resources [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self.driver.spawn(context, instance, image_meta, [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.683725] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] vm_ref = self.build_virtual_machine(instance, [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] for vif in network_info: [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return self._sync_wrapper(fn, *args, **kwargs) [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self.wait() [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self[:] = self._gt.wait() [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return self._exit_event.wait() [ 626.684138] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] result = hub.switch() [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return self.greenlet.switch() [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] result = function(*args, **kwargs) [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return func(*args, **kwargs) [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] raise e [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] nwinfo = self.network_api.allocate_for_instance( [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.684474] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] created_port_ids = self._update_ports_for_instance( [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] with excutils.save_and_reraise_exception(): [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self.force_reraise() [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] raise self.value [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] updated_port = self._update_port( [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] _ensure_no_port_binding_failure(port) [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.684786] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] raise exception.PortBindingFailed(port_id=port['id']) [ 626.685312] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 626.685312] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] [ 626.685312] env[62952]: INFO nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Terminating instance [ 626.690219] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquiring lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.690616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquired lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.690901] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.771508] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064177} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.771956] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.773142] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a79ef6d-960e-4596-b817-4415781c3644 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.793979] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.795022] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Successfully created port: 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.796883] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0427d5ca-ab31-44f3-aa72-1b22bc407936 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.819624] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 626.819624] env[62952]: value = "task-1367050" [ 626.819624] env[62952]: _type = "Task" [ 626.819624] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.829366] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.038782] env[62952]: DEBUG nova.network.neutron [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.093023] env[62952]: DEBUG nova.network.neutron [-] [instance: f569eee5-8806-4493-8625-52b55513f64c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.124070] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 627.151689] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:34:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='928944920',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-879848531',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 627.152659] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 627.152921] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.153147] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 627.153296] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.153440] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 627.153650] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 627.154972] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 627.154972] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 627.154972] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 627.154972] env[62952]: DEBUG nova.virt.hardware [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.155698] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.156335] env[62952]: ERROR nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Traceback (most recent call last): [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self.driver.spawn(context, instance, image_meta, [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] vm_ref = self.build_virtual_machine(instance, [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.156335] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] for vif in network_info: [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return self._sync_wrapper(fn, *args, **kwargs) [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self.wait() [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self[:] = self._gt.wait() [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return self._exit_event.wait() [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] result = hub.switch() [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.156722] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return self.greenlet.switch() [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] result = function(*args, **kwargs) [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] return func(*args, **kwargs) [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] raise e [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] nwinfo = self.network_api.allocate_for_instance( [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] created_port_ids = self._update_ports_for_instance( [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] with excutils.save_and_reraise_exception(): [ 627.157118] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] self.force_reraise() [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] raise self.value [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] updated_port = self._update_port( [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] _ensure_no_port_binding_failure(port) [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] raise exception.PortBindingFailed(port_id=port['id']) [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] nova.exception.PortBindingFailed: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. [ 627.157493] env[62952]: ERROR nova.compute.manager [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] [ 627.157817] env[62952]: DEBUG nova.compute.utils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.159464] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e111d59-6257-4a7c-b3bc-cd8955201743 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.163787] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Build of instance 21c8106e-8b6c-447b-89e2-df1b34aed12c was re-scheduled: Binding failed for port f17af0e2-a750-4c58-85f6-1028277f314d, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.164283] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.165866] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquiring lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.165866] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Acquired lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.165866] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.166094] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.675s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.171701] env[62952]: INFO nova.compute.claims [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.181641] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19fa105-6e1f-466b-bfe4-8d7c8306f310 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.296984] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.335281] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367050, 'name': ReconfigVM_Task, 'duration_secs': 0.257082} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.335660] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d/418cd947-00b8-485f-b9fc-2a675c58998d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.336515] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab5035d6-3525-4aa0-9f6d-b3688a653fbf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.343855] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 627.343855] env[62952]: value = "task-1367051" [ 627.343855] env[62952]: _type = "Task" [ 627.343855] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.356655] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367051, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.470355] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.500100] env[62952]: DEBUG nova.network.neutron [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] [instance: f569eee5-8806-4493-8625-52b55513f64c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.596612] env[62952]: INFO nova.compute.manager [-] [instance: f569eee5-8806-4493-8625-52b55513f64c] Took 1.05 seconds to deallocate network for instance. [ 627.601280] env[62952]: DEBUG nova.compute.claims [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.601838] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.758420] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.857458] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367051, 'name': Rename_Task, 'duration_secs': 0.136926} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.857458] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 627.857458] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-834466cc-4876-431c-908e-425b5c5e3b00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.864636] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Waiting for the task: (returnval){ [ 627.864636] env[62952]: value = "task-1367052" [ 627.864636] env[62952]: _type = "Task" [ 627.864636] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.876295] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367052, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.976710] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Releasing lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.977174] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 627.977629] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.978186] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aaf33838-c754-4dff-9c68-bc386e276ddc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.991758] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0f0c3c-163f-4bf2-bff0-e9918783f8c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.005236] env[62952]: DEBUG oslo_concurrency.lockutils [req-21e4172b-0221-458e-ad74-cb42e3440aa7 req-f5d4478b-9d15-410e-a1b1-9cb129281626 service nova] Releasing lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.018387] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f363ccb6-2c0e-4520-83b1-f41c74ab8bd3 could not be found. [ 628.018794] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.019207] env[62952]: INFO nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 628.019606] env[62952]: DEBUG oslo.service.loopingcall [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.019964] env[62952]: DEBUG nova.compute.manager [-] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.020197] env[62952]: DEBUG nova.network.neutron [-] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.066537] env[62952]: DEBUG nova.network.neutron [-] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.083185] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.259985] env[62952]: DEBUG nova.compute.manager [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f569eee5-8806-4493-8625-52b55513f64c] Received event network-vif-deleted-288e9672-8694-49b1-aea7-8532f1373dec {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.260231] env[62952]: DEBUG nova.compute.manager [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Received event network-changed-201ac188-f487-4348-97ee-c7533f56677f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.260389] env[62952]: DEBUG nova.compute.manager [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Refreshing instance network info cache due to event network-changed-201ac188-f487-4348-97ee-c7533f56677f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.260590] env[62952]: DEBUG oslo_concurrency.lockutils [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] Acquiring lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.260725] env[62952]: DEBUG oslo_concurrency.lockutils [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] Acquired lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.260869] env[62952]: DEBUG nova.network.neutron [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Refreshing network info cache for port 201ac188-f487-4348-97ee-c7533f56677f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.384014] env[62952]: DEBUG oslo_vmware.api [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Task: {'id': task-1367052, 'name': PowerOnVM_Task, 'duration_secs': 0.448334} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.384508] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 628.385095] env[62952]: DEBUG nova.compute.manager [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 628.385878] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208c0334-bac3-4541-ba9a-7708e9db1640 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.550573] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8f9285-65e6-4acb-9337-8815d432ec7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.563024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac1e719-a112-4771-ba0d-0a69c2cd287d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.569746] env[62952]: DEBUG nova.network.neutron [-] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.602658] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Releasing lock "refresh_cache-21c8106e-8b6c-447b-89e2-df1b34aed12c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.603012] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.603012] env[62952]: DEBUG nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.603160] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.606471] env[62952]: INFO nova.compute.manager [-] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Took 0.59 seconds to deallocate network for instance. [ 628.607413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e863fcc-854b-4f9c-a2b8-23aefc2f2ac4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.613293] env[62952]: DEBUG nova.compute.claims [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.613762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.619940] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4e59c0-c42d-4595-8ea7-d43bed1e2631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.640506] env[62952]: DEBUG nova.compute.provider_tree [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 628.641933] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.803970] env[62952]: DEBUG nova.network.neutron [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.911326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.146211] env[62952]: DEBUG nova.network.neutron [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.170768] env[62952]: ERROR nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [req-3d98a7c0-7692-4dcd-9789-c523e8a1ee6d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3d98a7c0-7692-4dcd-9789-c523e8a1ee6d"}]} [ 629.198779] env[62952]: DEBUG nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 629.215185] env[62952]: DEBUG nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 629.215441] env[62952]: DEBUG nova.compute.provider_tree [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 629.233033] env[62952]: DEBUG nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 629.258037] env[62952]: DEBUG nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 629.429019] env[62952]: DEBUG nova.network.neutron [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.654735] env[62952]: INFO nova.compute.manager [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] [instance: 21c8106e-8b6c-447b-89e2-df1b34aed12c] Took 1.05 seconds to deallocate network for instance. [ 629.684570] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645faa68-4486-4707-92f9-899358915633 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.700839] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01015f34-10ab-4878-90a5-67e890655240 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.737986] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd60f3c-0eeb-4813-8181-19342ea93a0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.747463] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b49a9f-9cf1-4309-9689-46a1757bb8ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.765020] env[62952]: DEBUG nova.compute.provider_tree [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 629.934968] env[62952]: DEBUG oslo_concurrency.lockutils [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] Releasing lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.936128] env[62952]: DEBUG nova.compute.manager [req-750d38d7-a54a-4754-987b-b19b32a7d49f req-3f0103e8-b9f9-4c1f-8d5e-bbc292d39944 service nova] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Received event network-vif-deleted-201ac188-f487-4348-97ee-c7533f56677f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.302949] env[62952]: DEBUG nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 29 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 630.303309] env[62952]: DEBUG nova.compute.provider_tree [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 29 to 30 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 630.303458] env[62952]: DEBUG nova.compute.provider_tree [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 630.444227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "418cd947-00b8-485f-b9fc-2a675c58998d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.444227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "418cd947-00b8-485f-b9fc-2a675c58998d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.444227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "418cd947-00b8-485f-b9fc-2a675c58998d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.444543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "418cd947-00b8-485f-b9fc-2a675c58998d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.444543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "418cd947-00b8-485f-b9fc-2a675c58998d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.447107] env[62952]: INFO nova.compute.manager [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Terminating instance [ 630.450160] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "refresh_cache-418cd947-00b8-485f-b9fc-2a675c58998d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.450328] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquired lock "refresh_cache-418cd947-00b8-485f-b9fc-2a675c58998d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.450496] env[62952]: DEBUG nova.network.neutron [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.710168] env[62952]: INFO nova.scheduler.client.report [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Deleted allocations for instance 21c8106e-8b6c-447b-89e2-df1b34aed12c [ 630.808929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.642s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.810042] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.816950] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.324s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.818563] env[62952]: INFO nova.compute.claims [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.995442] env[62952]: DEBUG nova.network.neutron [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.019252] env[62952]: ERROR nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 631.019252] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.019252] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.019252] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.019252] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.019252] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.019252] env[62952]: ERROR nova.compute.manager raise self.value [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.019252] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.019252] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.019252] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.020307] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.020307] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.020307] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 631.020307] env[62952]: ERROR nova.compute.manager [ 631.020307] env[62952]: Traceback (most recent call last): [ 631.020307] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.020307] env[62952]: listener.cb(fileno) [ 631.020307] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.020307] env[62952]: result = function(*args, **kwargs) [ 631.020307] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.020307] env[62952]: return func(*args, **kwargs) [ 631.020307] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.020307] env[62952]: raise e [ 631.020307] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.020307] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 631.020307] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.020307] env[62952]: created_port_ids = self._update_ports_for_instance( [ 631.020307] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.020307] env[62952]: with excutils.save_and_reraise_exception(): [ 631.020307] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.020307] env[62952]: self.force_reraise() [ 631.020307] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.020307] env[62952]: raise self.value [ 631.020307] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.020307] env[62952]: updated_port = self._update_port( [ 631.020307] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.020307] env[62952]: _ensure_no_port_binding_failure(port) [ 631.020307] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.020307] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.021808] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 631.021808] env[62952]: Removing descriptor: 19 [ 631.021808] env[62952]: ERROR nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Traceback (most recent call last): [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] yield resources [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self.driver.spawn(context, instance, image_meta, [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.021808] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] vm_ref = self.build_virtual_machine(instance, [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] for vif in network_info: [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return self._sync_wrapper(fn, *args, **kwargs) [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self.wait() [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self[:] = self._gt.wait() [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return self._exit_event.wait() [ 631.022378] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] result = hub.switch() [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return self.greenlet.switch() [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] result = function(*args, **kwargs) [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return func(*args, **kwargs) [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] raise e [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] nwinfo = self.network_api.allocate_for_instance( [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.023285] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] created_port_ids = self._update_ports_for_instance( [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] with excutils.save_and_reraise_exception(): [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self.force_reraise() [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] raise self.value [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] updated_port = self._update_port( [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] _ensure_no_port_binding_failure(port) [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.023991] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] raise exception.PortBindingFailed(port_id=port['id']) [ 631.024532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 631.024532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] [ 631.024532] env[62952]: INFO nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Terminating instance [ 631.027835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquiring lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.027987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquired lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.031733] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.076428] env[62952]: DEBUG nova.network.neutron [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.221952] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7038279d-b2e9-4cb1-86e1-3d201a572c49 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445 tempest-FloatingIPsAssociationNegativeTestJSON-1705724445-project-member] Lock "21c8106e-8b6c-447b-89e2-df1b34aed12c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.665s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.233599] env[62952]: DEBUG nova.compute.manager [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Received event network-changed-384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.233599] env[62952]: DEBUG nova.compute.manager [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Refreshing instance network info cache due to event network-changed-384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.233599] env[62952]: DEBUG oslo_concurrency.lockutils [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] Acquiring lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.327795] env[62952]: DEBUG nova.compute.utils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.334431] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.337116] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 631.475316] env[62952]: DEBUG nova.policy [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51631fa04d9c44eebb7aa61332193582', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab12fb45215443f1b10911673f10c192', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.583442] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Releasing lock "refresh_cache-418cd947-00b8-485f-b9fc-2a675c58998d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.584096] env[62952]: DEBUG nova.compute.manager [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.586018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.586356] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80369ed3-1d5c-49ba-9c4d-e03396cc433f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.599342] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 631.599679] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f1b8b16-235d-4b87-8682-27d3af37402e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.608055] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.613630] env[62952]: DEBUG oslo_vmware.api [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 631.613630] env[62952]: value = "task-1367055" [ 631.613630] env[62952]: _type = "Task" [ 631.613630] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.625643] env[62952]: DEBUG oslo_vmware.api [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.724594] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.835034] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.967368] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.133150] env[62952]: DEBUG oslo_vmware.api [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367055, 'name': PowerOffVM_Task, 'duration_secs': 0.129115} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.133685] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 632.133863] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 632.134148] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b8f345d-e36e-4e9e-87e3-734da3a3fa08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.161870] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cedd42d-560e-4c36-8faa-558bf07ca80e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.174188] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9e3847-8651-463e-9405-4796761a58ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.178163] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 632.178163] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 632.178430] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleting the datastore file [datastore2] 418cd947-00b8-485f-b9fc-2a675c58998d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 632.178557] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11d95892-0016-4177-a813-0366e4bcba5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.222922] env[62952]: DEBUG oslo_vmware.api [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for the task: (returnval){ [ 632.222922] env[62952]: value = "task-1367057" [ 632.222922] env[62952]: _type = "Task" [ 632.222922] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.223766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5b092f-ac7e-4037-b13a-32f669aad4c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.239628] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c334ed2-8268-4c43-a2fc-a51d17a93c11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.243932] env[62952]: DEBUG oslo_vmware.api [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.260114] env[62952]: DEBUG nova.compute.provider_tree [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.262349] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.471544] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Releasing lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.472591] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.475017] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.475017] env[62952]: DEBUG oslo_concurrency.lockutils [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] Acquired lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.475017] env[62952]: DEBUG nova.network.neutron [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Refreshing network info cache for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.475017] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d6aebd9-9c9d-4ab1-8077-aaf72e8d0330 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.486055] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ae29aa-b7f8-4644-ab3e-b185c7e86d64 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.516889] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51db1590-9612-4d23-a49f-e1900f82ed90 could not be found. [ 632.519428] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.519428] env[62952]: INFO nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Took 0.05 seconds to destroy the instance on the hypervisor. [ 632.519428] env[62952]: DEBUG oslo.service.loopingcall [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.519428] env[62952]: DEBUG nova.compute.manager [-] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.519428] env[62952]: DEBUG nova.network.neutron [-] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.536606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquiring lock "cb35ba14-6a78-4381-9c93-f2fa285468d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.536606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Lock "cb35ba14-6a78-4381-9c93-f2fa285468d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.592892] env[62952]: DEBUG nova.network.neutron [-] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.737723] env[62952]: DEBUG oslo_vmware.api [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Task: {'id': task-1367057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108197} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.737723] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 632.738123] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 632.738123] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.738272] env[62952]: INFO nova.compute.manager [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 632.738569] env[62952]: DEBUG oslo.service.loopingcall [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.738735] env[62952]: DEBUG nova.compute.manager [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.738783] env[62952]: DEBUG nova.network.neutron [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.764189] env[62952]: DEBUG nova.scheduler.client.report [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.775704] env[62952]: DEBUG nova.network.neutron [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.848596] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.876751] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.876996] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.877194] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.877380] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.877526] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.877672] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.878229] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.878428] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.878604] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.878767] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.878937] env[62952]: DEBUG nova.virt.hardware [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.879838] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7f1778-d4b8-4ed6-9055-e77ac493b39e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.888785] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635d9ce7-a5bd-4f7c-a6f4-8fc422b3cb1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.904436] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Successfully created port: 02b395c8-1824-407f-8d77-060a295fe5e5 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.009250] env[62952]: DEBUG nova.network.neutron [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.098650] env[62952]: DEBUG nova.network.neutron [-] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.269662] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.270741] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.273592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.320s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.278525] env[62952]: INFO nova.compute.claims [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.284037] env[62952]: DEBUG nova.network.neutron [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.292315] env[62952]: DEBUG nova.network.neutron [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.601150] env[62952]: INFO nova.compute.manager [-] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Took 1.08 seconds to deallocate network for instance. [ 633.604791] env[62952]: DEBUG nova.compute.claims [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.605739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.784446] env[62952]: DEBUG nova.compute.utils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.790283] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.790478] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.793206] env[62952]: INFO nova.compute.manager [-] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Took 1.05 seconds to deallocate network for instance. [ 633.797751] env[62952]: DEBUG oslo_concurrency.lockutils [req-f58f9221-cbe7-4b27-a7ec-30bdfcef79ce req-d16d33ef-5b1d-4ca7-bc68-03eb908b5bf8 service nova] Releasing lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.911633] env[62952]: DEBUG nova.policy [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '460f4570c2074153a6c0e4ae06e4d99a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bde4ba5200d04548b3a1893afc56720a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.922814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquiring lock "64044aea-4572-48b1-93b3-93f079a829dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.923636] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Lock "64044aea-4572-48b1-93b3-93f079a829dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.290773] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.303120] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.634085] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9649c54f-7415-4f69-b1be-1459a7f8b5cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.649043] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93fdb461-aa5d-42ac-a0cd-892c84ae4e9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.691360] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2f224b-9d7e-47cf-b986-3bb6db764a1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.701446] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f355df4e-969a-40f8-b4b0-a7aedde0a397 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.718335] env[62952]: DEBUG nova.compute.provider_tree [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 634.789828] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Successfully created port: 312162a5-494e-4f64-b28a-659c3a7bb039 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.817388] env[62952]: DEBUG nova.compute.manager [req-790d7af9-a05d-4a5d-8e1a-cd82e70d9471 req-3ea4c6d8-4b0a-43d3-a66d-1b34f1cd551f service nova] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Received event network-vif-deleted-384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.255202] env[62952]: ERROR nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [req-e6af5f56-9e88-4225-87a9-d3d08aab4336] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e6af5f56-9e88-4225-87a9-d3d08aab4336"}]} [ 635.280079] env[62952]: DEBUG nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 635.298556] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.302684] env[62952]: DEBUG nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 635.302916] env[62952]: DEBUG nova.compute.provider_tree [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 635.330680] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.330962] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.331145] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.331243] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.331441] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.331546] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.332135] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.332135] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.333311] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.333537] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.333745] env[62952]: DEBUG nova.virt.hardware [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.334700] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27151433-f93d-4d1a-9692-4bcbf681c625 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.338510] env[62952]: DEBUG nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 635.351477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435f96af-2e98-494d-88b8-11a322ec9f6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.379652] env[62952]: DEBUG nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 635.753025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2b0d35-4d38-47d6-8a0c-eef16c04e7cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.761976] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef9d780-5f4e-485b-825c-fc6ccd2abbff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.801662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608eecc2-4fca-4126-aa79-985158567a00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.811160] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eae4dac-1013-4b99-8fc9-ef6b2dbc6c2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.827464] env[62952]: DEBUG nova.compute.provider_tree [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 635.959292] env[62952]: ERROR nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 635.959292] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.959292] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.959292] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.959292] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.959292] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.959292] env[62952]: ERROR nova.compute.manager raise self.value [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.959292] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.959292] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.959292] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.960189] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.960189] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.960189] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 635.960189] env[62952]: ERROR nova.compute.manager [ 635.960189] env[62952]: Traceback (most recent call last): [ 635.960189] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.960189] env[62952]: listener.cb(fileno) [ 635.960189] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.960189] env[62952]: result = function(*args, **kwargs) [ 635.960189] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.960189] env[62952]: return func(*args, **kwargs) [ 635.960189] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.960189] env[62952]: raise e [ 635.960189] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.960189] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 635.960189] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.960189] env[62952]: created_port_ids = self._update_ports_for_instance( [ 635.960189] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.960189] env[62952]: with excutils.save_and_reraise_exception(): [ 635.960189] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.960189] env[62952]: self.force_reraise() [ 635.960189] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.960189] env[62952]: raise self.value [ 635.960189] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.960189] env[62952]: updated_port = self._update_port( [ 635.960189] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.960189] env[62952]: _ensure_no_port_binding_failure(port) [ 635.960189] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.960189] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.960974] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 635.960974] env[62952]: Removing descriptor: 15 [ 635.960974] env[62952]: ERROR nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Traceback (most recent call last): [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] yield resources [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self.driver.spawn(context, instance, image_meta, [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.960974] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] vm_ref = self.build_virtual_machine(instance, [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] for vif in network_info: [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return self._sync_wrapper(fn, *args, **kwargs) [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self.wait() [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self[:] = self._gt.wait() [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return self._exit_event.wait() [ 635.961362] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] result = hub.switch() [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return self.greenlet.switch() [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] result = function(*args, **kwargs) [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return func(*args, **kwargs) [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] raise e [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] nwinfo = self.network_api.allocate_for_instance( [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.961782] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] created_port_ids = self._update_ports_for_instance( [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] with excutils.save_and_reraise_exception(): [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self.force_reraise() [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] raise self.value [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] updated_port = self._update_port( [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] _ensure_no_port_binding_failure(port) [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.962237] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] raise exception.PortBindingFailed(port_id=port['id']) [ 635.962608] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 635.962608] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] [ 635.962608] env[62952]: INFO nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Terminating instance [ 635.962608] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquiring lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.962608] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquired lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.962608] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.372122] env[62952]: DEBUG nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 34 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 636.372122] env[62952]: DEBUG nova.compute.provider_tree [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 34 to 35 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 636.372258] env[62952]: DEBUG nova.compute.provider_tree [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 636.514327] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.653263] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.877884] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.604s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.878780] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.881790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.436s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.910668] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.915673] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.005s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.156231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Releasing lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.156795] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.156945] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.157342] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be029ec5-1471-454c-9bff-5b5450421cbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.172122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a488597-ae29-44c7-a79f-88609cda0252 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.206203] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1dc0311e-2f03-402a-bb31-117289502ab0 could not be found. [ 637.206665] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.207075] env[62952]: INFO nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 637.207189] env[62952]: DEBUG oslo.service.loopingcall [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.207410] env[62952]: DEBUG nova.compute.manager [-] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.207505] env[62952]: DEBUG nova.network.neutron [-] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.229570] env[62952]: DEBUG nova.network.neutron [-] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.360756] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquiring lock "710a4cb2-a1f6-461e-ac26-d1e86cdd435d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.361448] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "710a4cb2-a1f6-461e-ac26-d1e86cdd435d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.387269] env[62952]: DEBUG nova.compute.utils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.393213] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 637.393414] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 637.566666] env[62952]: DEBUG nova.policy [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b752d9b13c50423297fa0c2336f7638a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1dd5a879b4e42e98e871e4739c0f279', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.733021] env[62952]: DEBUG nova.network.neutron [-] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.800800] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f06b5cf-4c1a-4db0-bf3f-16274052b5d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.815631] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfe5267-b614-47a2-8ab1-9547e9b6345e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.821585] env[62952]: ERROR nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 637.821585] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.821585] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.821585] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.821585] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.821585] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.821585] env[62952]: ERROR nova.compute.manager raise self.value [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.821585] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.821585] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.821585] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.822233] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.822233] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.822233] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 637.822233] env[62952]: ERROR nova.compute.manager [ 637.822233] env[62952]: Traceback (most recent call last): [ 637.822233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.822233] env[62952]: listener.cb(fileno) [ 637.822233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.822233] env[62952]: result = function(*args, **kwargs) [ 637.822233] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.822233] env[62952]: return func(*args, **kwargs) [ 637.822233] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.822233] env[62952]: raise e [ 637.822233] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.822233] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 637.822233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.822233] env[62952]: created_port_ids = self._update_ports_for_instance( [ 637.822233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.822233] env[62952]: with excutils.save_and_reraise_exception(): [ 637.822233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.822233] env[62952]: self.force_reraise() [ 637.822233] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.822233] env[62952]: raise self.value [ 637.822233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.822233] env[62952]: updated_port = self._update_port( [ 637.822233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.822233] env[62952]: _ensure_no_port_binding_failure(port) [ 637.822233] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.822233] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.823140] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 637.823140] env[62952]: Removing descriptor: 19 [ 637.823140] env[62952]: ERROR nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Traceback (most recent call last): [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] yield resources [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self.driver.spawn(context, instance, image_meta, [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.823140] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] vm_ref = self.build_virtual_machine(instance, [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] for vif in network_info: [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return self._sync_wrapper(fn, *args, **kwargs) [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self.wait() [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self[:] = self._gt.wait() [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return self._exit_event.wait() [ 637.823517] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] result = hub.switch() [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return self.greenlet.switch() [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] result = function(*args, **kwargs) [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return func(*args, **kwargs) [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] raise e [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] nwinfo = self.network_api.allocate_for_instance( [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.823945] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] created_port_ids = self._update_ports_for_instance( [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] with excutils.save_and_reraise_exception(): [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self.force_reraise() [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] raise self.value [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] updated_port = self._update_port( [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] _ensure_no_port_binding_failure(port) [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.824340] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] raise exception.PortBindingFailed(port_id=port['id']) [ 637.824774] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 637.824774] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] [ 637.824774] env[62952]: INFO nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Terminating instance [ 637.830296] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquiring lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.830406] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquired lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.830735] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.861635] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fc1ae9-5822-4db4-8b8e-bb1c129f8db2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.870400] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeaf9631-2aa8-4cc6-826d-d36aa9f2a9f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.888316] env[62952]: DEBUG nova.compute.provider_tree [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.893944] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.911770] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.927860] env[62952]: DEBUG nova.compute.manager [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Received event network-changed-02b395c8-1824-407f-8d77-060a295fe5e5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.928703] env[62952]: DEBUG nova.compute.manager [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Refreshing instance network info cache due to event network-changed-02b395c8-1824-407f-8d77-060a295fe5e5. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.930349] env[62952]: DEBUG oslo_concurrency.lockutils [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] Acquiring lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.930349] env[62952]: DEBUG oslo_concurrency.lockutils [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] Acquired lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.930349] env[62952]: DEBUG nova.network.neutron [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Refreshing network info cache for port 02b395c8-1824-407f-8d77-060a295fe5e5 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.173865] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.235397] env[62952]: INFO nova.compute.manager [-] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Took 1.03 seconds to deallocate network for instance. [ 638.238880] env[62952]: DEBUG nova.compute.claims [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.239059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.388105] env[62952]: DEBUG nova.scheduler.client.report [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.488245] env[62952]: DEBUG nova.network.neutron [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.660163] env[62952]: DEBUG nova.network.neutron [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.675519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Releasing lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.675961] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.676178] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.676761] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d897ba1-cbef-43ae-93c3-60245a310201 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.690640] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ec5526-04eb-4b77-a5c2-8622b1afdb88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.720422] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6511f90d-82e8-4374-bdaf-0a6cc66a5a65 could not be found. [ 638.720422] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.720422] env[62952]: INFO nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Took 0.04 seconds to destroy the instance on the hypervisor. [ 638.720422] env[62952]: DEBUG oslo.service.loopingcall [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.720698] env[62952]: DEBUG nova.compute.manager [-] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.720734] env[62952]: DEBUG nova.network.neutron [-] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.771716] env[62952]: DEBUG nova.network.neutron [-] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.863999] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Successfully created port: 5deee2ee-241b-4ebe-b06c-7e2c19331b14 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.882747] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "de2daf8c-73c1-44fe-bb01-2b54f1f1c419" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.882999] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "de2daf8c-73c1-44fe-bb01-2b54f1f1c419" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.895307] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.895914] env[62952]: ERROR nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] Traceback (most recent call last): [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self.driver.spawn(context, instance, image_meta, [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] vm_ref = self.build_virtual_machine(instance, [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.895914] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] for vif in network_info: [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return self._sync_wrapper(fn, *args, **kwargs) [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self.wait() [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self[:] = self._gt.wait() [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return self._exit_event.wait() [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] result = hub.switch() [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.896495] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return self.greenlet.switch() [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] result = function(*args, **kwargs) [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] return func(*args, **kwargs) [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] raise e [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] nwinfo = self.network_api.allocate_for_instance( [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] created_port_ids = self._update_ports_for_instance( [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] with excutils.save_and_reraise_exception(): [ 638.897438] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] self.force_reraise() [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] raise self.value [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] updated_port = self._update_port( [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] _ensure_no_port_binding_failure(port) [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] raise exception.PortBindingFailed(port_id=port['id']) [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] nova.exception.PortBindingFailed: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. [ 638.898169] env[62952]: ERROR nova.compute.manager [instance: 77418633-2833-47a7-b30f-dca84432d382] [ 638.898699] env[62952]: DEBUG nova.compute.utils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.899154] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Build of instance 77418633-2833-47a7-b30f-dca84432d382 was re-scheduled: Binding failed for port 126d35b6-a8f6-4147-9b91-7ca5da49a4df, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.899533] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.899763] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.899913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquired lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.900087] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.901825] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.800s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.903300] env[62952]: INFO nova.compute.claims [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.913128] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.948068] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.948239] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.948587] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.948587] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.948700] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.948826] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.949044] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.949186] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.949372] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.949521] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.949820] env[62952]: DEBUG nova.virt.hardware [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.950590] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d036ecc-c157-4ad0-a147-a3404ad6b355 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.961104] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8954dd-b95a-436b-8e40-8c88373dab8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.163871] env[62952]: DEBUG oslo_concurrency.lockutils [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] Releasing lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.164587] env[62952]: DEBUG nova.compute.manager [req-deb51784-d973-4e3e-a15f-bb7dc4474acb req-472d10c5-8739-4bd4-b0bc-a1a7d034e60a service nova] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Received event network-vif-deleted-02b395c8-1824-407f-8d77-060a295fe5e5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.275058] env[62952]: DEBUG nova.network.neutron [-] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.385688] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "244c71ad-130d-44ad-9998-3f5c51518482" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.385688] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "244c71ad-130d-44ad-9998-3f5c51518482" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.444596] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.663837] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.745870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquiring lock "a224dd44-bda2-464f-b946-edbe167927b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.747142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Lock "a224dd44-bda2-464f-b946-edbe167927b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.778453] env[62952]: INFO nova.compute.manager [-] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Took 1.06 seconds to deallocate network for instance. [ 639.781482] env[62952]: DEBUG nova.compute.claims [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.781482] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.166107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Releasing lock "refresh_cache-77418633-2833-47a7-b30f-dca84432d382" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.166535] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.166604] env[62952]: DEBUG nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.167122] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.216893] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.342501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd76c91-e6f1-4775-9b2c-edef2f6cbefc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.351856] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbe9cad-7b4d-41d1-930f-3b9884e2a7d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.394165] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2beb43c-09fb-4736-966a-cd30c8de2b3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.402527] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecc5796-ea3a-452d-8bdd-3ccb8052ba8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.419535] env[62952]: DEBUG nova.compute.provider_tree [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.721244] env[62952]: DEBUG nova.network.neutron [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.923263] env[62952]: DEBUG nova.scheduler.client.report [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.961847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquiring lock "869c5e7f-b937-4add-ace2-7d9b178c2acc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.962324] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Lock "869c5e7f-b937-4add-ace2-7d9b178c2acc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.154505] env[62952]: DEBUG nova.compute.manager [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Received event network-changed-312162a5-494e-4f64-b28a-659c3a7bb039 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.154811] env[62952]: DEBUG nova.compute.manager [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Refreshing instance network info cache due to event network-changed-312162a5-494e-4f64-b28a-659c3a7bb039. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.155159] env[62952]: DEBUG oslo_concurrency.lockutils [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] Acquiring lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.155359] env[62952]: DEBUG oslo_concurrency.lockutils [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] Acquired lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.155572] env[62952]: DEBUG nova.network.neutron [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Refreshing network info cache for port 312162a5-494e-4f64-b28a-659c3a7bb039 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.224580] env[62952]: INFO nova.compute.manager [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 77418633-2833-47a7-b30f-dca84432d382] Took 1.06 seconds to deallocate network for instance. [ 641.429895] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.433997] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 641.435289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.330s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.435660] env[62952]: DEBUG nova.objects.instance [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 641.465914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "ac386ddb-c812-4689-b9b4-b98d26ad1831" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.466218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "ac386ddb-c812-4689-b9b4-b98d26ad1831" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.706118] env[62952]: DEBUG nova.network.neutron [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.941057] env[62952]: DEBUG nova.compute.utils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 641.956023] env[62952]: DEBUG nova.network.neutron [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.958481] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.963020] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 641.963020] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 642.014847] env[62952]: ERROR nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 642.014847] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.014847] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.014847] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.014847] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.014847] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.014847] env[62952]: ERROR nova.compute.manager raise self.value [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.014847] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.014847] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.014847] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.015400] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.015400] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.015400] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 642.015400] env[62952]: ERROR nova.compute.manager [ 642.015400] env[62952]: Traceback (most recent call last): [ 642.015400] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.015400] env[62952]: listener.cb(fileno) [ 642.015400] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.015400] env[62952]: result = function(*args, **kwargs) [ 642.015400] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.015400] env[62952]: return func(*args, **kwargs) [ 642.015400] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.015400] env[62952]: raise e [ 642.015400] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.015400] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 642.015400] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.015400] env[62952]: created_port_ids = self._update_ports_for_instance( [ 642.015400] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.015400] env[62952]: with excutils.save_and_reraise_exception(): [ 642.015400] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.015400] env[62952]: self.force_reraise() [ 642.015400] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.015400] env[62952]: raise self.value [ 642.015400] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.015400] env[62952]: updated_port = self._update_port( [ 642.015400] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.015400] env[62952]: _ensure_no_port_binding_failure(port) [ 642.015400] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.015400] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.016256] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 642.016256] env[62952]: Removing descriptor: 15 [ 642.016256] env[62952]: ERROR nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Traceback (most recent call last): [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] yield resources [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self.driver.spawn(context, instance, image_meta, [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.016256] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] vm_ref = self.build_virtual_machine(instance, [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] for vif in network_info: [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return self._sync_wrapper(fn, *args, **kwargs) [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self.wait() [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self[:] = self._gt.wait() [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return self._exit_event.wait() [ 642.016621] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] result = hub.switch() [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return self.greenlet.switch() [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] result = function(*args, **kwargs) [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return func(*args, **kwargs) [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] raise e [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] nwinfo = self.network_api.allocate_for_instance( [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.016993] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] created_port_ids = self._update_ports_for_instance( [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] with excutils.save_and_reraise_exception(): [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self.force_reraise() [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] raise self.value [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] updated_port = self._update_port( [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] _ensure_no_port_binding_failure(port) [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.017379] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] raise exception.PortBindingFailed(port_id=port['id']) [ 642.017744] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 642.017744] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] [ 642.017744] env[62952]: INFO nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Terminating instance [ 642.021606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquiring lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.021950] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquired lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.021950] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.123965] env[62952]: DEBUG nova.policy [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d254e604e82453d95f85247c49898a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '602039593ae34a56a98329d98d94eb5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 642.284086] env[62952]: INFO nova.scheduler.client.report [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Deleted allocations for instance 77418633-2833-47a7-b30f-dca84432d382 [ 642.470856] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d103b9ca-0bea-48b6-aa34-5b70c9181be8 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.036s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.474317] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.990s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.478021] env[62952]: INFO nova.compute.claims [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.479826] env[62952]: DEBUG oslo_concurrency.lockutils [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] Releasing lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.479984] env[62952]: DEBUG nova.compute.manager [req-51ae74eb-af91-4bed-8d3e-992cdf3187f0 req-36902db8-7710-4df0-8a24-6ffadec213d2 service nova] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Received event network-vif-deleted-312162a5-494e-4f64-b28a-659c3a7bb039 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 642.580805] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.798357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d1d26ed7-b320-4401-b677-155f116cdace tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "77418633-2833-47a7-b30f-dca84432d382" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.630s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.815992] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.974538] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 643.015115] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Successfully created port: 4fdfc6d8-f6af-4171-9c89-4937b1d59604 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.022143] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 643.022694] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 643.022694] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.022815] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 643.022852] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.023029] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 643.023243] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 643.023395] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 643.023552] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 643.023751] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 643.023869] env[62952]: DEBUG nova.virt.hardware [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.025116] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51732bd0-b4ae-4d7c-9743-3ba372dea419 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.035684] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45aa238d-a820-4c74-9cd0-b301a9f62b09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.303611] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.320129] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Releasing lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.320753] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.321063] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.323183] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c011ff21-c374-4292-a070-17caba608eaa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.336452] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d485c11f-304c-43c8-a973-5d1d28552d2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.375117] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 722f0334-6210-46b8-87de-d9bf24812a1f could not be found. [ 643.375117] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.375117] env[62952]: INFO nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 643.375117] env[62952]: DEBUG oslo.service.loopingcall [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.375117] env[62952]: DEBUG nova.compute.manager [-] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.375117] env[62952]: DEBUG nova.network.neutron [-] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.419305] env[62952]: DEBUG nova.network.neutron [-] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.498808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "f4ec1e94-1778-4296-989d-782700c6faf8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.499065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "f4ec1e94-1778-4296-989d-782700c6faf8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.838972] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.919227] env[62952]: DEBUG nova.network.neutron [-] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.019342] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808e34a0-d22b-476a-a05e-2ea37f183e70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.026759] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464321ea-0d79-4a21-a44f-09329ceaaebe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.058478] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e6649b-ce7e-4acc-a329-4d1f933b2ed9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.068180] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db281b59-9d7f-4c48-acb5-38aab7f77457 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.083183] env[62952]: DEBUG nova.compute.provider_tree [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.238814] env[62952]: DEBUG nova.compute.manager [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Received event network-changed-5deee2ee-241b-4ebe-b06c-7e2c19331b14 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.239088] env[62952]: DEBUG nova.compute.manager [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Refreshing instance network info cache due to event network-changed-5deee2ee-241b-4ebe-b06c-7e2c19331b14. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 644.239375] env[62952]: DEBUG oslo_concurrency.lockutils [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] Acquiring lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.239553] env[62952]: DEBUG oslo_concurrency.lockutils [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] Acquired lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.239807] env[62952]: DEBUG nova.network.neutron [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Refreshing network info cache for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 644.427974] env[62952]: INFO nova.compute.manager [-] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Took 1.05 seconds to deallocate network for instance. [ 644.429269] env[62952]: DEBUG nova.compute.claims [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.429694] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.588858] env[62952]: DEBUG nova.scheduler.client.report [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.766415] env[62952]: DEBUG nova.network.neutron [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.898658] env[62952]: DEBUG nova.network.neutron [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.099898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.627s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.100410] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.105303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.503s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.207561] env[62952]: ERROR nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 645.207561] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.207561] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.207561] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.207561] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.207561] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.207561] env[62952]: ERROR nova.compute.manager raise self.value [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.207561] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.207561] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.207561] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.210180] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.210180] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.210180] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 645.210180] env[62952]: ERROR nova.compute.manager [ 645.210180] env[62952]: Traceback (most recent call last): [ 645.210180] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.210180] env[62952]: listener.cb(fileno) [ 645.210180] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.210180] env[62952]: result = function(*args, **kwargs) [ 645.210180] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.210180] env[62952]: return func(*args, **kwargs) [ 645.210180] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.210180] env[62952]: raise e [ 645.210180] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.210180] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 645.210180] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.210180] env[62952]: created_port_ids = self._update_ports_for_instance( [ 645.210180] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.210180] env[62952]: with excutils.save_and_reraise_exception(): [ 645.210180] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.210180] env[62952]: self.force_reraise() [ 645.210180] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.210180] env[62952]: raise self.value [ 645.210180] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.210180] env[62952]: updated_port = self._update_port( [ 645.210180] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.210180] env[62952]: _ensure_no_port_binding_failure(port) [ 645.210180] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.210180] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.211081] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 645.211081] env[62952]: Removing descriptor: 19 [ 645.211081] env[62952]: ERROR nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] Traceback (most recent call last): [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] yield resources [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self.driver.spawn(context, instance, image_meta, [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.211081] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] vm_ref = self.build_virtual_machine(instance, [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] for vif in network_info: [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return self._sync_wrapper(fn, *args, **kwargs) [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self.wait() [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self[:] = self._gt.wait() [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return self._exit_event.wait() [ 645.211457] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] result = hub.switch() [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return self.greenlet.switch() [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] result = function(*args, **kwargs) [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return func(*args, **kwargs) [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] raise e [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] nwinfo = self.network_api.allocate_for_instance( [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.211867] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] created_port_ids = self._update_ports_for_instance( [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] with excutils.save_and_reraise_exception(): [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self.force_reraise() [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] raise self.value [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] updated_port = self._update_port( [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] _ensure_no_port_binding_failure(port) [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.212294] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] raise exception.PortBindingFailed(port_id=port['id']) [ 645.212662] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 645.212662] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] [ 645.212662] env[62952]: INFO nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Terminating instance [ 645.213734] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.213861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquired lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.214135] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.321676] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.321754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.357792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "a825d324-0b26-4051-bdfa-4a4246f6d4e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.358166] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "a825d324-0b26-4051-bdfa-4a4246f6d4e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.382409] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.382617] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.400880] env[62952]: DEBUG oslo_concurrency.lockutils [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] Releasing lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.401150] env[62952]: DEBUG nova.compute.manager [req-b36d43de-f84d-4612-b9ea-4936431305ee req-2d915f93-8aea-402f-b9e2-50478ea44aab service nova] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Received event network-vif-deleted-5deee2ee-241b-4ebe-b06c-7e2c19331b14 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.613807] env[62952]: DEBUG nova.compute.utils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.622834] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.623103] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.778031] env[62952]: DEBUG nova.policy [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88d8d97572264d6a99d58e9e34fdf956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1427ec86cfc64c1e8db03445199a4d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 645.780331] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.051409] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.126481] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.157848] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e935577a-5ab3-4834-b131-0fd50d0dd05a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.168558] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f17a93-6f01-4d87-94f5-ecf3dcc98096 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.209036] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fa3c17-d679-4c59-bdef-f5ac67f68fe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.217455] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7440711e-e5c9-4f35-8b37-4e51a18208d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.233589] env[62952]: DEBUG nova.compute.provider_tree [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.554492] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Releasing lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.554919] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 646.555122] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 646.555433] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-008073ff-e3c5-4784-9668-afdf9cfeaf9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.567200] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872c749e-6fc6-46bb-8f6f-d1ade7c9cb3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.583933] env[62952]: DEBUG nova.compute.manager [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Received event network-changed-4fdfc6d8-f6af-4171-9c89-4937b1d59604 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.584284] env[62952]: DEBUG nova.compute.manager [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Refreshing instance network info cache due to event network-changed-4fdfc6d8-f6af-4171-9c89-4937b1d59604. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.584678] env[62952]: DEBUG oslo_concurrency.lockutils [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] Acquiring lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.584905] env[62952]: DEBUG oslo_concurrency.lockutils [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] Acquired lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.587034] env[62952]: DEBUG nova.network.neutron [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Refreshing network info cache for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.595420] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 510731f2-399d-486a-8c60-1c61c05063de could not be found. [ 646.595614] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 646.595822] env[62952]: INFO nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Took 0.04 seconds to destroy the instance on the hypervisor. [ 646.596230] env[62952]: DEBUG oslo.service.loopingcall [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.596996] env[62952]: DEBUG nova.compute.manager [-] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.597203] env[62952]: DEBUG nova.network.neutron [-] [instance: 510731f2-399d-486a-8c60-1c61c05063de] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.654102] env[62952]: DEBUG nova.network.neutron [-] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.667705] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Successfully created port: 12c4f6c4-1948-436d-a327-1d6785e7dc5b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.740303] env[62952]: DEBUG nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.134123] env[62952]: DEBUG nova.network.neutron [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.145089] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.157594] env[62952]: DEBUG nova.network.neutron [-] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.176040] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.176040] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.176762] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.176983] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.177567] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.177567] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.177567] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.177742] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.177824] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.177998] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.178185] env[62952]: DEBUG nova.virt.hardware [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.179374] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037927d9-42ce-4fdb-9d78-b4bdf2b7d5ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.192407] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efeb1210-ceca-41da-b2d1-8e66b485d5c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.245189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.140s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.247590] env[62952]: ERROR nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] Traceback (most recent call last): [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self.driver.spawn(context, instance, image_meta, [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] vm_ref = self.build_virtual_machine(instance, [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.247590] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] for vif in network_info: [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return self._sync_wrapper(fn, *args, **kwargs) [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self.wait() [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self[:] = self._gt.wait() [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return self._exit_event.wait() [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] result = hub.switch() [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.247993] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return self.greenlet.switch() [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] result = function(*args, **kwargs) [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] return func(*args, **kwargs) [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] raise e [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] nwinfo = self.network_api.allocate_for_instance( [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] created_port_ids = self._update_ports_for_instance( [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] with excutils.save_and_reraise_exception(): [ 647.248400] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] self.force_reraise() [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] raise self.value [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] updated_port = self._update_port( [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] _ensure_no_port_binding_failure(port) [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] raise exception.PortBindingFailed(port_id=port['id']) [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] nova.exception.PortBindingFailed: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. [ 647.248835] env[62952]: ERROR nova.compute.manager [instance: f569eee5-8806-4493-8625-52b55513f64c] [ 647.249184] env[62952]: DEBUG nova.compute.utils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 647.253018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.637s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.253916] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Build of instance f569eee5-8806-4493-8625-52b55513f64c was re-scheduled: Binding failed for port 288e9672-8694-49b1-aea7-8532f1373dec, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 647.254906] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 647.255513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquiring lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.255790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Acquired lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.256095] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.328519] env[62952]: DEBUG nova.network.neutron [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.662083] env[62952]: INFO nova.compute.manager [-] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Took 1.06 seconds to deallocate network for instance. [ 647.665334] env[62952]: DEBUG nova.compute.claims [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.665334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.798735] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.832270] env[62952]: DEBUG oslo_concurrency.lockutils [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] Releasing lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.832519] env[62952]: DEBUG nova.compute.manager [req-9e5644aa-f61f-4b21-b03c-cc0156ccbca9 req-7b9dbf53-c0be-4b20-b5e9-37c9a9232b76 service nova] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Received event network-vif-deleted-4fdfc6d8-f6af-4171-9c89-4937b1d59604 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.031663] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.227325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquiring lock "9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.227427] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Lock "9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.252977] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76d3b09-22e0-46a9-9015-2a600f181916 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.265566] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d6cf35-62b6-4e59-8984-c355fbbc8829 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.303050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7de2ac-df16-48b4-bf36-d9f0c86e6702 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.314145] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49121ef9-9492-4a50-8482-328d7ca159cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.330337] env[62952]: DEBUG nova.compute.provider_tree [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.537022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Releasing lock "refresh_cache-f569eee5-8806-4493-8625-52b55513f64c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.537022] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 648.537022] env[62952]: DEBUG nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.537022] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.571802] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.736374] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquiring lock "38e05fe2-d938-43d6-ab2b-b99fb7be51d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.736654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Lock "38e05fe2-d938-43d6-ab2b-b99fb7be51d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.785881] env[62952]: ERROR nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 648.785881] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.785881] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.785881] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.785881] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.785881] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.785881] env[62952]: ERROR nova.compute.manager raise self.value [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.785881] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.785881] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.785881] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.786427] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.786427] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.786427] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 648.786427] env[62952]: ERROR nova.compute.manager [ 648.786427] env[62952]: Traceback (most recent call last): [ 648.786427] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.786427] env[62952]: listener.cb(fileno) [ 648.786427] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.786427] env[62952]: result = function(*args, **kwargs) [ 648.786427] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.786427] env[62952]: return func(*args, **kwargs) [ 648.786427] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.786427] env[62952]: raise e [ 648.786427] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.786427] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 648.786427] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.786427] env[62952]: created_port_ids = self._update_ports_for_instance( [ 648.786427] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.786427] env[62952]: with excutils.save_and_reraise_exception(): [ 648.786427] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.786427] env[62952]: self.force_reraise() [ 648.786427] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.786427] env[62952]: raise self.value [ 648.786427] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.786427] env[62952]: updated_port = self._update_port( [ 648.786427] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.786427] env[62952]: _ensure_no_port_binding_failure(port) [ 648.786427] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.786427] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.787306] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 648.787306] env[62952]: Removing descriptor: 19 [ 648.787306] env[62952]: ERROR nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] Traceback (most recent call last): [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] yield resources [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self.driver.spawn(context, instance, image_meta, [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.787306] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] vm_ref = self.build_virtual_machine(instance, [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] for vif in network_info: [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return self._sync_wrapper(fn, *args, **kwargs) [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self.wait() [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self[:] = self._gt.wait() [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return self._exit_event.wait() [ 648.787778] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] result = hub.switch() [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return self.greenlet.switch() [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] result = function(*args, **kwargs) [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return func(*args, **kwargs) [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] raise e [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] nwinfo = self.network_api.allocate_for_instance( [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.788233] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] created_port_ids = self._update_ports_for_instance( [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] with excutils.save_and_reraise_exception(): [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self.force_reraise() [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] raise self.value [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] updated_port = self._update_port( [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] _ensure_no_port_binding_failure(port) [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.788660] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] raise exception.PortBindingFailed(port_id=port['id']) [ 648.792760] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 648.792760] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] [ 648.792760] env[62952]: INFO nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Terminating instance [ 648.792760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquiring lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.792760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquired lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.792760] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.835920] env[62952]: DEBUG nova.scheduler.client.report [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.075974] env[62952]: DEBUG nova.network.neutron [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.248886] env[62952]: DEBUG nova.compute.manager [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] [instance: 7c98a241-9297-417a-b412-5c279144572b] Received event network-changed-12c4f6c4-1948-436d-a327-1d6785e7dc5b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.249037] env[62952]: DEBUG nova.compute.manager [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] [instance: 7c98a241-9297-417a-b412-5c279144572b] Refreshing instance network info cache due to event network-changed-12c4f6c4-1948-436d-a327-1d6785e7dc5b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 649.249646] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] Acquiring lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.317605] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.340745] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.341968] env[62952]: ERROR nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Traceback (most recent call last): [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self.driver.spawn(context, instance, image_meta, [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] vm_ref = self.build_virtual_machine(instance, [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.341968] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] for vif in network_info: [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return self._sync_wrapper(fn, *args, **kwargs) [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self.wait() [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self[:] = self._gt.wait() [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return self._exit_event.wait() [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] result = hub.switch() [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.342345] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return self.greenlet.switch() [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] result = function(*args, **kwargs) [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] return func(*args, **kwargs) [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] raise e [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] nwinfo = self.network_api.allocate_for_instance( [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] created_port_ids = self._update_ports_for_instance( [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] with excutils.save_and_reraise_exception(): [ 649.342707] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] self.force_reraise() [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] raise self.value [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] updated_port = self._update_port( [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] _ensure_no_port_binding_failure(port) [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] raise exception.PortBindingFailed(port_id=port['id']) [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] nova.exception.PortBindingFailed: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. [ 649.343121] env[62952]: ERROR nova.compute.manager [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] [ 649.343435] env[62952]: DEBUG nova.compute.utils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 649.343892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.433s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.344249] env[62952]: DEBUG nova.objects.instance [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 649.346683] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Build of instance f363ccb6-2c0e-4520-83b1-f41c74ab8bd3 was re-scheduled: Binding failed for port 201ac188-f487-4348-97ee-c7533f56677f, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 649.347521] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 649.347521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquiring lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.347521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Acquired lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.347521] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.408635] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.582760] env[62952]: INFO nova.compute.manager [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] [instance: f569eee5-8806-4493-8625-52b55513f64c] Took 1.05 seconds to deallocate network for instance. [ 649.885795] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.901845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "5283bf79-bf20-4a38-912f-ced401fbd691" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.902105] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "5283bf79-bf20-4a38-912f-ced401fbd691" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.911372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Releasing lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.911773] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 649.911978] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.912307] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] Acquired lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.912482] env[62952]: DEBUG nova.network.neutron [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] [instance: 7c98a241-9297-417a-b412-5c279144572b] Refreshing network info cache for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 649.913464] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05ca9df2-19cb-4b6c-8c1b-1b01cde1d8c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.924059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fa9b01-955e-4e19-9695-4a00c69a223c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.948962] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c98a241-9297-417a-b412-5c279144572b could not be found. [ 649.949258] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.949407] env[62952]: INFO nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 649.949661] env[62952]: DEBUG oslo.service.loopingcall [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.951840] env[62952]: DEBUG nova.compute.manager [-] [instance: 7c98a241-9297-417a-b412-5c279144572b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.951940] env[62952]: DEBUG nova.network.neutron [-] [instance: 7c98a241-9297-417a-b412-5c279144572b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.980219] env[62952]: DEBUG nova.network.neutron [-] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.014260] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.360956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d08c10d9-7902-4be2-a6e6-9d3a0194f328 tempest-ServersAdmin275Test-1373400290 tempest-ServersAdmin275Test-1373400290-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.362082] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.100s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.365142] env[62952]: INFO nova.compute.claims [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.451631] env[62952]: DEBUG nova.network.neutron [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.482594] env[62952]: DEBUG nova.network.neutron [-] [instance: 7c98a241-9297-417a-b412-5c279144572b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.517113] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Releasing lock "refresh_cache-f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.517424] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 650.517682] env[62952]: DEBUG nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.517864] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.537373] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.585632] env[62952]: DEBUG nova.network.neutron [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] [instance: 7c98a241-9297-417a-b412-5c279144572b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.629516] env[62952]: INFO nova.scheduler.client.report [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Deleted allocations for instance f569eee5-8806-4493-8625-52b55513f64c [ 650.985412] env[62952]: INFO nova.compute.manager [-] [instance: 7c98a241-9297-417a-b412-5c279144572b] Took 1.03 seconds to deallocate network for instance. [ 650.992768] env[62952]: DEBUG nova.compute.claims [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 650.993533] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.041322] env[62952]: DEBUG nova.network.neutron [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.088845] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9c3ff20-30f2-477e-b24c-9309e06be110 req-7b3e332b-0e8a-48ba-9f7d-934269cc6a47 service nova] Releasing lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.137218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fe7fc4d3-3c36-45be-868b-082636f0fc2b tempest-VolumesAssistedSnapshotsTest-586058241 tempest-VolumesAssistedSnapshotsTest-586058241-project-member] Lock "f569eee5-8806-4493-8625-52b55513f64c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.442825] env[62952]: DEBUG nova.compute.manager [req-c79c428f-0b21-4f94-88c7-8cf802ae2541 req-abd1ef91-0aa0-450f-b70a-b0c5fc5fa786 service nova] [instance: 7c98a241-9297-417a-b412-5c279144572b] Received event network-vif-deleted-12c4f6c4-1948-436d-a327-1d6785e7dc5b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.547676] env[62952]: INFO nova.compute.manager [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] [instance: f363ccb6-2c0e-4520-83b1-f41c74ab8bd3] Took 1.03 seconds to deallocate network for instance. [ 651.643184] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.925238] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3c4061-019c-4baf-a78b-14534c09ef29 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.934403] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f66088-ab52-46b1-9186-0dac3019445d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.973134] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974717a5-3065-410a-9add-9add5c57e547 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.982718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41972cb-74bf-4326-bb07-9617fda25053 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.996815] env[62952]: DEBUG nova.compute.provider_tree [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.164984] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.502892] env[62952]: DEBUG nova.scheduler.client.report [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.544096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquiring lock "fc5b891b-8432-4b2f-97d9-c93ca0a18783" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.544205] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Lock "fc5b891b-8432-4b2f-97d9-c93ca0a18783" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.585276] env[62952]: INFO nova.scheduler.client.report [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Deleted allocations for instance f363ccb6-2c0e-4520-83b1-f41c74ab8bd3 [ 653.009761] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.646s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.009761] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.015852] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.411s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.096084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2f70cfe3-656c-499f-b84d-405e5e2a3df9 tempest-ServerAddressesTestJSON-881028171 tempest-ServerAddressesTestJSON-881028171-project-member] Lock "f363ccb6-2c0e-4520-83b1-f41c74ab8bd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.662s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.521081] env[62952]: DEBUG nova.compute.utils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.522477] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.522648] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.599045] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 653.625782] env[62952]: DEBUG nova.policy [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb47d681e0354b47b431e9a717bc9e05', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cb3a1ee49b64d5db965dc072f290577', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.031505] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.096651] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04632816-0416-4b0f-a640-fcdadc0ded5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.109794] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7107de2-3be4-498b-8c3b-73d73e2ce82c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.146269] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.147102] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a6c5d2-e27c-46f2-812f-1d0c5b575d35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.155622] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6763027f-f6f0-4d16-8869-050ad51b488c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.170838] env[62952]: DEBUG nova.compute.provider_tree [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.612898] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Successfully created port: 85839fde-aea9-4fe2-8bad-808ff10fcc08 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.678151] env[62952]: DEBUG nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.042772] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.085048] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.085288] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.085442] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.085621] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.085766] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.085915] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.086381] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.086651] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.086804] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.086947] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.087157] env[62952]: DEBUG nova.virt.hardware [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.088655] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e46310-2b78-41e3-acc9-3cf1b3f3e1c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.098706] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478cbeff-6991-4f94-8f9d-bc1c3511ae84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.183182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.166s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.183182] env[62952]: ERROR nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Traceback (most recent call last): [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self.driver.spawn(context, instance, image_meta, [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.183182] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] vm_ref = self.build_virtual_machine(instance, [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] for vif in network_info: [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return self._sync_wrapper(fn, *args, **kwargs) [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self.wait() [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self[:] = self._gt.wait() [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return self._exit_event.wait() [ 655.183532] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] result = hub.switch() [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return self.greenlet.switch() [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] result = function(*args, **kwargs) [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] return func(*args, **kwargs) [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] raise e [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] nwinfo = self.network_api.allocate_for_instance( [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.184028] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] created_port_ids = self._update_ports_for_instance( [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] with excutils.save_and_reraise_exception(): [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] self.force_reraise() [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] raise self.value [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] updated_port = self._update_port( [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] _ensure_no_port_binding_failure(port) [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.184654] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] raise exception.PortBindingFailed(port_id=port['id']) [ 655.185978] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] nova.exception.PortBindingFailed: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. [ 655.185978] env[62952]: ERROR nova.compute.manager [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] [ 655.185978] env[62952]: DEBUG nova.compute.utils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.185978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.882s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.185978] env[62952]: DEBUG nova.objects.instance [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lazy-loading 'resources' on Instance uuid 418cd947-00b8-485f-b9fc-2a675c58998d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 655.189057] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Build of instance 51db1590-9612-4d23-a49f-e1900f82ed90 was re-scheduled: Binding failed for port 384cb6d0-5dd5-4d90-80b1-6dfaaf5bbb56, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.189057] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.189057] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquiring lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.189057] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Acquired lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.189284] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.733309] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.001256] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.231986] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9555cc-63fe-44af-a29a-4b4a6d942ac8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.239913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13153a95-30c2-46aa-b674-192c5e264cba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.275581] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45a4389-b5d9-4f20-b1ce-1ade5deb9c99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.286975] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3879cd-8fc2-43db-852a-4525c6e89c9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.301880] env[62952]: DEBUG nova.compute.provider_tree [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.419495] env[62952]: DEBUG nova.compute.manager [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Received event network-changed-85839fde-aea9-4fe2-8bad-808ff10fcc08 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.419713] env[62952]: DEBUG nova.compute.manager [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Refreshing instance network info cache due to event network-changed-85839fde-aea9-4fe2-8bad-808ff10fcc08. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.419912] env[62952]: DEBUG oslo_concurrency.lockutils [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] Acquiring lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.420037] env[62952]: DEBUG oslo_concurrency.lockutils [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] Acquired lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.420201] env[62952]: DEBUG nova.network.neutron [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Refreshing network info cache for port 85839fde-aea9-4fe2-8bad-808ff10fcc08 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 656.507017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Releasing lock "refresh_cache-51db1590-9612-4d23-a49f-e1900f82ed90" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.507017] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.507017] env[62952]: DEBUG nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.507017] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.542210] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.804972] env[62952]: DEBUG nova.scheduler.client.report [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.825740] env[62952]: ERROR nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 656.825740] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.825740] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.825740] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.825740] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.825740] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.825740] env[62952]: ERROR nova.compute.manager raise self.value [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.825740] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.825740] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.825740] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.826498] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.826498] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.826498] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 656.826498] env[62952]: ERROR nova.compute.manager [ 656.826498] env[62952]: Traceback (most recent call last): [ 656.826498] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.826498] env[62952]: listener.cb(fileno) [ 656.826498] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.826498] env[62952]: result = function(*args, **kwargs) [ 656.826498] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.826498] env[62952]: return func(*args, **kwargs) [ 656.826498] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.826498] env[62952]: raise e [ 656.826498] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.826498] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 656.826498] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.826498] env[62952]: created_port_ids = self._update_ports_for_instance( [ 656.826498] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.826498] env[62952]: with excutils.save_and_reraise_exception(): [ 656.826498] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.826498] env[62952]: self.force_reraise() [ 656.826498] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.826498] env[62952]: raise self.value [ 656.826498] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.826498] env[62952]: updated_port = self._update_port( [ 656.826498] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.826498] env[62952]: _ensure_no_port_binding_failure(port) [ 656.826498] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.826498] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.827901] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 656.827901] env[62952]: Removing descriptor: 19 [ 656.827901] env[62952]: ERROR nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Traceback (most recent call last): [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] yield resources [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self.driver.spawn(context, instance, image_meta, [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.827901] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] vm_ref = self.build_virtual_machine(instance, [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] for vif in network_info: [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return self._sync_wrapper(fn, *args, **kwargs) [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self.wait() [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self[:] = self._gt.wait() [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return self._exit_event.wait() [ 656.828480] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] result = hub.switch() [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return self.greenlet.switch() [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] result = function(*args, **kwargs) [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return func(*args, **kwargs) [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] raise e [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] nwinfo = self.network_api.allocate_for_instance( [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.828918] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] created_port_ids = self._update_ports_for_instance( [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] with excutils.save_and_reraise_exception(): [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self.force_reraise() [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] raise self.value [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] updated_port = self._update_port( [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] _ensure_no_port_binding_failure(port) [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.829328] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] raise exception.PortBindingFailed(port_id=port['id']) [ 656.829666] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 656.829666] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] [ 656.829666] env[62952]: INFO nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Terminating instance [ 656.830222] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquiring lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.043121] env[62952]: DEBUG nova.network.neutron [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.095514] env[62952]: DEBUG nova.network.neutron [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.201098] env[62952]: DEBUG nova.network.neutron [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.310322] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.313320] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.074s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.334873] env[62952]: INFO nova.scheduler.client.report [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Deleted allocations for instance 418cd947-00b8-485f-b9fc-2a675c58998d [ 657.546314] env[62952]: INFO nova.compute.manager [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] [instance: 51db1590-9612-4d23-a49f-e1900f82ed90] Took 1.04 seconds to deallocate network for instance. [ 657.706395] env[62952]: DEBUG oslo_concurrency.lockutils [req-2de6e6cb-92b4-45fe-be6a-e0fc194b98b1 req-8d79b535-1b3c-457f-ae93-6f53450c82f6 service nova] Releasing lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.706715] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "9567865d-41ae-4379-8313-e87bfc92fba8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.710019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "9567865d-41ae-4379-8313-e87bfc92fba8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.710019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquired lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.710019] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.847564] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eae15fdb-68af-4fb7-94ce-4585a0b40346 tempest-ServersAdmin275Test-900092893 tempest-ServersAdmin275Test-900092893-project-member] Lock "418cd947-00b8-485f-b9fc-2a675c58998d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.403s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.236909] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.339450] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.388153] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdf8756-1f2d-42fe-bda8-2f9f3dc6a7e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.397911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008bbb01-7ca7-49d6-96bf-0cb0ef0b7ad1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.435903] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b737e31e-225e-4a9d-814c-bf86363407a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.444696] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75f835a-80e8-4c0f-aa2b-b6da57bc86a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.462158] env[62952]: DEBUG nova.compute.provider_tree [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.480290] env[62952]: DEBUG nova.compute.manager [req-7aeba2d0-9b29-4487-890e-4117c1523d42 req-42481bcf-6927-48fe-8b49-2b77662b0cc0 service nova] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Received event network-vif-deleted-85839fde-aea9-4fe2-8bad-808ff10fcc08 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.587896] env[62952]: INFO nova.scheduler.client.report [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Deleted allocations for instance 51db1590-9612-4d23-a49f-e1900f82ed90 [ 658.842533] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Releasing lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.843875] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.843875] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.843875] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ecbf2d7-5a7a-4da9-a2f5-46f3617e7eca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.853366] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b00593-2a75-4918-b620-4e8ab8f721ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.877283] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 542169bf-71e2-47f8-a311-8a3dca4ffee7 could not be found. [ 658.877518] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.877708] env[62952]: INFO nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 658.877959] env[62952]: DEBUG oslo.service.loopingcall [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.878213] env[62952]: DEBUG nova.compute.manager [-] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.878297] env[62952]: DEBUG nova.network.neutron [-] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.901835] env[62952]: DEBUG nova.network.neutron [-] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.967311] env[62952]: DEBUG nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.097938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e5c315e1-ceba-4a53-bd15-e4fd267910ef tempest-ServersWithSpecificFlavorTestJSON-1407891863 tempest-ServersWithSpecificFlavorTestJSON-1407891863-project-member] Lock "51db1590-9612-4d23-a49f-e1900f82ed90" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.316s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.403220] env[62952]: DEBUG nova.network.neutron [-] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.471732] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.159s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.472370] env[62952]: ERROR nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Traceback (most recent call last): [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self.driver.spawn(context, instance, image_meta, [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] vm_ref = self.build_virtual_machine(instance, [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.472370] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] for vif in network_info: [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return self._sync_wrapper(fn, *args, **kwargs) [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self.wait() [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self[:] = self._gt.wait() [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return self._exit_event.wait() [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] result = hub.switch() [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.472825] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return self.greenlet.switch() [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] result = function(*args, **kwargs) [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] return func(*args, **kwargs) [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] raise e [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] nwinfo = self.network_api.allocate_for_instance( [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] created_port_ids = self._update_ports_for_instance( [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] with excutils.save_and_reraise_exception(): [ 659.474404] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] self.force_reraise() [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] raise self.value [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] updated_port = self._update_port( [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] _ensure_no_port_binding_failure(port) [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] raise exception.PortBindingFailed(port_id=port['id']) [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] nova.exception.PortBindingFailed: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. [ 659.475442] env[62952]: ERROR nova.compute.manager [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] [ 659.477022] env[62952]: DEBUG nova.compute.utils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.477022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.694s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.478053] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Build of instance 1dc0311e-2f03-402a-bb31-117289502ab0 was re-scheduled: Binding failed for port 02b395c8-1824-407f-8d77-060a295fe5e5, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.478288] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.478515] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquiring lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.478737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Acquired lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.479028] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.600589] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.907015] env[62952]: INFO nova.compute.manager [-] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Took 1.03 seconds to deallocate network for instance. [ 659.911738] env[62952]: DEBUG nova.compute.claims [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.912650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.014117] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.096571] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.136343] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.440161] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3dfca2-5b67-43fe-ad6b-2b379bb8e9cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.449022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0469e6dd-5bad-43ad-87c2-53f127ef41db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.481404] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664b5897-48e8-4aa4-a705-215557cb8756 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.489950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b20bf41-ec4f-4f20-925d-6054ee98e672 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.504290] env[62952]: DEBUG nova.compute.provider_tree [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.601225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Releasing lock "refresh_cache-1dc0311e-2f03-402a-bb31-117289502ab0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.601473] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.601638] env[62952]: DEBUG nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.601796] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.617795] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.008329] env[62952]: DEBUG nova.scheduler.client.report [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.123733] env[62952]: DEBUG nova.network.neutron [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.515552] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.041s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.516209] env[62952]: ERROR nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Traceback (most recent call last): [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self.driver.spawn(context, instance, image_meta, [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] vm_ref = self.build_virtual_machine(instance, [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.516209] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] for vif in network_info: [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return self._sync_wrapper(fn, *args, **kwargs) [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self.wait() [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self[:] = self._gt.wait() [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return self._exit_event.wait() [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] result = hub.switch() [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.516589] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return self.greenlet.switch() [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] result = function(*args, **kwargs) [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] return func(*args, **kwargs) [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] raise e [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] nwinfo = self.network_api.allocate_for_instance( [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] created_port_ids = self._update_ports_for_instance( [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] with excutils.save_and_reraise_exception(): [ 661.516963] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] self.force_reraise() [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] raise self.value [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] updated_port = self._update_port( [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] _ensure_no_port_binding_failure(port) [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] raise exception.PortBindingFailed(port_id=port['id']) [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] nova.exception.PortBindingFailed: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. [ 661.517503] env[62952]: ERROR nova.compute.manager [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] [ 661.517837] env[62952]: DEBUG nova.compute.utils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.521742] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.683s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.525712] env[62952]: INFO nova.compute.claims [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.533009] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Build of instance 6511f90d-82e8-4374-bdaf-0a6cc66a5a65 was re-scheduled: Binding failed for port 312162a5-494e-4f64-b28a-659c3a7bb039, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.533541] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.533721] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquiring lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.533899] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Acquired lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.536144] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.627540] env[62952]: INFO nova.compute.manager [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] [instance: 1dc0311e-2f03-402a-bb31-117289502ab0] Took 1.02 seconds to deallocate network for instance. [ 662.059107] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.156438] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.660357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Releasing lock "refresh_cache-6511f90d-82e8-4374-bdaf-0a6cc66a5a65" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.660703] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.660833] env[62952]: DEBUG nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.661758] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.664112] env[62952]: INFO nova.scheduler.client.report [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Deleted allocations for instance 1dc0311e-2f03-402a-bb31-117289502ab0 [ 662.710196] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.070019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5641af95-a905-4085-b282-2d3da54dd375 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.078790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd56b400-0baf-49bb-885d-124de2ac4f4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.117763] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae1b812-8896-4b9f-b0ed-194abef7131a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.131373] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294d4f12-1a0a-4191-8286-05131ef63fe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.147885] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.175223] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6738522-06ce-40a4-9e6d-ee5b5ece0dfa tempest-ServerExternalEventsTest-1176729700 tempest-ServerExternalEventsTest-1176729700-project-member] Lock "1dc0311e-2f03-402a-bb31-117289502ab0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.733s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.212957] env[62952]: DEBUG nova.network.neutron [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.653355] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.677788] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.716232] env[62952]: INFO nova.compute.manager [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] [instance: 6511f90d-82e8-4374-bdaf-0a6cc66a5a65] Took 1.05 seconds to deallocate network for instance. [ 663.909435] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "0f8162d7-c5e6-4c66-8a10-bb06998738b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.909861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "0f8162d7-c5e6-4c66-8a10-bb06998738b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.161326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.639s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.161934] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.169538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.736s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.212980] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.673085] env[62952]: DEBUG nova.compute.utils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.679609] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.680051] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 664.764261] env[62952]: DEBUG nova.policy [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d254e604e82453d95f85247c49898a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '602039593ae34a56a98329d98d94eb5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 664.791688] env[62952]: INFO nova.scheduler.client.report [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Deleted allocations for instance 6511f90d-82e8-4374-bdaf-0a6cc66a5a65 [ 664.896328] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "65c1a63f-4645-40fe-b429-00923e60d1cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.896519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "65c1a63f-4645-40fe-b429-00923e60d1cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.186026] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.273471] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff650a1-aa84-43b8-9be4-7d822ca5df8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.283122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcadc1aa-c5d9-457b-86bf-bd097785c12a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.320124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d247ef63-4303-404f-89a9-717cd7803b70 tempest-ServerPasswordTestJSON-1685617635 tempest-ServerPasswordTestJSON-1685617635-project-member] Lock "6511f90d-82e8-4374-bdaf-0a6cc66a5a65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.867s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.321906] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa67265-b8a8-46e0-a90f-b0e1a27bfaeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.329595] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79994c45-9284-4a56-a513-d9c9545b7229 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.343611] env[62952]: DEBUG nova.compute.provider_tree [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.553179] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Successfully created port: ba13b6c5-0256-4df0-b394-7d07b2970c0d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.826113] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.846702] env[62952]: DEBUG nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.203107] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.229785] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.230066] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.230237] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.230425] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.230572] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.230717] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.230923] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.231345] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.231568] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.231806] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.231999] env[62952]: DEBUG nova.virt.hardware [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.233111] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b2b755-5ef6-47fb-858c-414de0c26b44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.243397] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fce2942-15af-4fea-9dd0-faa535dea40d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.353554] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.186s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.353554] env[62952]: ERROR nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Traceback (most recent call last): [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self.driver.spawn(context, instance, image_meta, [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.353554] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] vm_ref = self.build_virtual_machine(instance, [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] for vif in network_info: [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return self._sync_wrapper(fn, *args, **kwargs) [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self.wait() [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self[:] = self._gt.wait() [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return self._exit_event.wait() [ 666.353767] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] result = hub.switch() [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return self.greenlet.switch() [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] result = function(*args, **kwargs) [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] return func(*args, **kwargs) [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] raise e [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] nwinfo = self.network_api.allocate_for_instance( [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.354045] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] created_port_ids = self._update_ports_for_instance( [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] with excutils.save_and_reraise_exception(): [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] self.force_reraise() [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] raise self.value [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] updated_port = self._update_port( [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] _ensure_no_port_binding_failure(port) [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.354376] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] raise exception.PortBindingFailed(port_id=port['id']) [ 666.354612] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] nova.exception.PortBindingFailed: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. [ 666.354612] env[62952]: ERROR nova.compute.manager [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] [ 666.354612] env[62952]: DEBUG nova.compute.utils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.356244] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.690s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.364841] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.364841] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Build of instance 722f0334-6210-46b8-87de-d9bf24812a1f was re-scheduled: Binding failed for port 5deee2ee-241b-4ebe-b06c-7e2c19331b14, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.364841] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.364841] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquiring lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.365297] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Acquired lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.365297] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.908392] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.932946] env[62952]: DEBUG nova.compute.manager [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Received event network-changed-ba13b6c5-0256-4df0-b394-7d07b2970c0d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.933233] env[62952]: DEBUG nova.compute.manager [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Refreshing instance network info cache due to event network-changed-ba13b6c5-0256-4df0-b394-7d07b2970c0d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 666.933472] env[62952]: DEBUG oslo_concurrency.lockutils [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] Acquiring lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.933550] env[62952]: DEBUG oslo_concurrency.lockutils [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] Acquired lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.933630] env[62952]: DEBUG nova.network.neutron [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Refreshing network info cache for port ba13b6c5-0256-4df0-b394-7d07b2970c0d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 667.007989] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.159169] env[62952]: ERROR nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 667.159169] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.159169] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.159169] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.159169] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.159169] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.159169] env[62952]: ERROR nova.compute.manager raise self.value [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.159169] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.159169] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.159169] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.159544] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.159544] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.159544] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 667.159544] env[62952]: ERROR nova.compute.manager [ 667.159544] env[62952]: Traceback (most recent call last): [ 667.159544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.159544] env[62952]: listener.cb(fileno) [ 667.159544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.159544] env[62952]: result = function(*args, **kwargs) [ 667.159544] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 667.159544] env[62952]: return func(*args, **kwargs) [ 667.159544] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.159544] env[62952]: raise e [ 667.159544] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.159544] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 667.159544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.159544] env[62952]: created_port_ids = self._update_ports_for_instance( [ 667.159544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.159544] env[62952]: with excutils.save_and_reraise_exception(): [ 667.159544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.159544] env[62952]: self.force_reraise() [ 667.159544] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.159544] env[62952]: raise self.value [ 667.159544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.159544] env[62952]: updated_port = self._update_port( [ 667.159544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.159544] env[62952]: _ensure_no_port_binding_failure(port) [ 667.159544] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.159544] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.160541] env[62952]: nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 667.160541] env[62952]: Removing descriptor: 19 [ 667.160541] env[62952]: ERROR nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Traceback (most recent call last): [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] yield resources [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self.driver.spawn(context, instance, image_meta, [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.160541] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] vm_ref = self.build_virtual_machine(instance, [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] for vif in network_info: [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return self._sync_wrapper(fn, *args, **kwargs) [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self.wait() [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self[:] = self._gt.wait() [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return self._exit_event.wait() [ 667.160816] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] result = hub.switch() [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return self.greenlet.switch() [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] result = function(*args, **kwargs) [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return func(*args, **kwargs) [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] raise e [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] nwinfo = self.network_api.allocate_for_instance( [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.161077] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] created_port_ids = self._update_ports_for_instance( [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] with excutils.save_and_reraise_exception(): [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self.force_reraise() [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] raise self.value [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] updated_port = self._update_port( [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] _ensure_no_port_binding_failure(port) [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.161331] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] raise exception.PortBindingFailed(port_id=port['id']) [ 667.161611] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 667.161611] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] [ 667.161611] env[62952]: INFO nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Terminating instance [ 667.166170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.383718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0eafdd-32a7-4a74-ab07-d8c49fb58d10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.398352] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61682ee1-b6b5-432b-9677-552d3a409a5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.435627] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3603f89a-248c-4d80-9f0a-77b4012f1b16 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.447026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145657ea-9228-49ad-8a9e-d20090ad13be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.460543] env[62952]: DEBUG nova.compute.provider_tree [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.462431] env[62952]: DEBUG nova.network.neutron [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.513327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Releasing lock "refresh_cache-722f0334-6210-46b8-87de-d9bf24812a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.513556] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.513721] env[62952]: DEBUG nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.513882] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.534423] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.569690] env[62952]: DEBUG nova.network.neutron [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.966157] env[62952]: DEBUG nova.scheduler.client.report [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.036596] env[62952]: DEBUG nova.network.neutron [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.073343] env[62952]: DEBUG oslo_concurrency.lockutils [req-3e0cbbe0-87c4-479f-a66a-a9b3857dff35 req-6f895889-c663-46b7-bd82-404cb226ac8a service nova] Releasing lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.073723] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquired lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.073947] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.179991] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 668.179991] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 668.474274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.118s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.474326] env[62952]: ERROR nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] Traceback (most recent call last): [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self.driver.spawn(context, instance, image_meta, [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] vm_ref = self.build_virtual_machine(instance, [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.474326] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] for vif in network_info: [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return self._sync_wrapper(fn, *args, **kwargs) [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self.wait() [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self[:] = self._gt.wait() [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return self._exit_event.wait() [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] result = hub.switch() [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.474558] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return self.greenlet.switch() [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] result = function(*args, **kwargs) [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] return func(*args, **kwargs) [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] raise e [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] nwinfo = self.network_api.allocate_for_instance( [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] created_port_ids = self._update_ports_for_instance( [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] with excutils.save_and_reraise_exception(): [ 668.474810] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] self.force_reraise() [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] raise self.value [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] updated_port = self._update_port( [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] _ensure_no_port_binding_failure(port) [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] raise exception.PortBindingFailed(port_id=port['id']) [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] nova.exception.PortBindingFailed: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. [ 668.475088] env[62952]: ERROR nova.compute.manager [instance: 510731f2-399d-486a-8c60-1c61c05063de] [ 668.475313] env[62952]: DEBUG nova.compute.utils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.477494] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Build of instance 510731f2-399d-486a-8c60-1c61c05063de was re-scheduled: Binding failed for port 4fdfc6d8-f6af-4171-9c89-4937b1d59604, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.477937] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.478182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.478324] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquired lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.478478] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.479532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.486s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.536170] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.538655] env[62952]: INFO nova.compute.manager [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] [instance: 722f0334-6210-46b8-87de-d9bf24812a1f] Took 1.02 seconds to deallocate network for instance. [ 668.611745] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.690132] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 668.690132] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 668.692323] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 668.806368] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.808708] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.998757] env[62952]: DEBUG nova.compute.manager [req-3482f448-ae43-49b8-95fa-46dcf74f9a71 req-bf882871-ff87-4462-84b4-6685cecb024c service nova] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Received event network-vif-deleted-ba13b6c5-0256-4df0-b394-7d07b2970c0d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.196411] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 669.196411] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 669.196411] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 669.196677] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.196916] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.197092] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.197148] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.197294] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.197475] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.197583] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 669.197740] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 669.307644] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Releasing lock "refresh_cache-510731f2-399d-486a-8c60-1c61c05063de" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.307897] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.308094] env[62952]: DEBUG nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.308261] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.310130] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Releasing lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.310477] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 669.310649] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.310931] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-782c86e7-2a65-49d6-ade5-80bc6c73bba7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.329150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e086787f-6946-47bb-a05c-2f0e5c8d214b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.337297] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.354019] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 12408f00-ed1d-4b96-bc0c-59779d0eda74 could not be found. [ 669.354019] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.356265] env[62952]: INFO nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Took 0.04 seconds to destroy the instance on the hypervisor. [ 669.356265] env[62952]: DEBUG oslo.service.loopingcall [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.356265] env[62952]: DEBUG nova.compute.manager [-] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.356265] env[62952]: DEBUG nova.network.neutron [-] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.385013] env[62952]: DEBUG nova.network.neutron [-] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.520063] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ba1ee3-641f-4e89-9268-352f826aeb1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.528250] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7ee005-b47b-4057-a1ff-a6a84a54d848 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.563783] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eaa03df-a124-444a-af37-eb2c8d00ac89 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.571766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee027e26-973c-49a7-b273-896a2166631c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.585785] env[62952]: DEBUG nova.compute.provider_tree [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.587898] env[62952]: INFO nova.scheduler.client.report [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Deleted allocations for instance 722f0334-6210-46b8-87de-d9bf24812a1f [ 669.702209] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.847113] env[62952]: DEBUG nova.network.neutron [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.893437] env[62952]: DEBUG nova.network.neutron [-] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.098434] env[62952]: DEBUG nova.scheduler.client.report [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.103176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ce08993a-4a08-4a94-8372-d7b5e60671ab tempest-ServersAdminNegativeTestJSON-162250365 tempest-ServersAdminNegativeTestJSON-162250365-project-member] Lock "722f0334-6210-46b8-87de-d9bf24812a1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.191s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.356027] env[62952]: INFO nova.compute.manager [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 510731f2-399d-486a-8c60-1c61c05063de] Took 1.05 seconds to deallocate network for instance. [ 670.398032] env[62952]: INFO nova.compute.manager [-] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Took 1.04 seconds to deallocate network for instance. [ 670.401208] env[62952]: DEBUG nova.compute.claims [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.401450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.608281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.129s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.608963] env[62952]: ERROR nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] Traceback (most recent call last): [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self.driver.spawn(context, instance, image_meta, [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] vm_ref = self.build_virtual_machine(instance, [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.608963] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] for vif in network_info: [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return self._sync_wrapper(fn, *args, **kwargs) [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self.wait() [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self[:] = self._gt.wait() [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return self._exit_event.wait() [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] result = hub.switch() [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.609271] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return self.greenlet.switch() [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] result = function(*args, **kwargs) [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] return func(*args, **kwargs) [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] raise e [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] nwinfo = self.network_api.allocate_for_instance( [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] created_port_ids = self._update_ports_for_instance( [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] with excutils.save_and_reraise_exception(): [ 670.609551] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] self.force_reraise() [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] raise self.value [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] updated_port = self._update_port( [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] _ensure_no_port_binding_failure(port) [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] raise exception.PortBindingFailed(port_id=port['id']) [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] nova.exception.PortBindingFailed: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. [ 670.609852] env[62952]: ERROR nova.compute.manager [instance: 7c98a241-9297-417a-b412-5c279144572b] [ 670.610099] env[62952]: DEBUG nova.compute.utils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.610946] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.616142] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Build of instance 7c98a241-9297-417a-b412-5c279144572b was re-scheduled: Binding failed for port 12c4f6c4-1948-436d-a327-1d6785e7dc5b, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.616142] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.616142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquiring lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.616142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Acquired lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.616383] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.617117] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.452s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.618677] env[62952]: INFO nova.compute.claims [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.004874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquiring lock "2222309b-ce4e-46eb-a620-34553f644947" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.007743] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Lock "2222309b-ce4e-46eb-a620-34553f644947" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.136877] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.147785] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.242125] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.385029] env[62952]: INFO nova.scheduler.client.report [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Deleted allocations for instance 510731f2-399d-486a-8c60-1c61c05063de [ 671.522590] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquiring lock "74361d48-0a82-453a-a4a1-45173c49e901" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.524077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Lock "74361d48-0a82-453a-a4a1-45173c49e901" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.745337] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Releasing lock "refresh_cache-7c98a241-9297-417a-b412-5c279144572b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.745592] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.745777] env[62952]: DEBUG nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.745944] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.775672] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.896360] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d04a7b68-d58d-436c-afb8-f78c976b0094 tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "510731f2-399d-486a-8c60-1c61c05063de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.492s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.147975] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08440005-86a7-423a-98ac-0526126beeea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.158098] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b64c86f-b280-44c7-a33d-c151ed384005 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.192685] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acc58e3-d70b-45bd-ae8e-e8b4db794f45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.200841] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7d1929-24ab-4d8d-8056-1420fcac0846 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.217070] env[62952]: DEBUG nova.compute.provider_tree [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.280515] env[62952]: DEBUG nova.network.neutron [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.400221] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.719728] env[62952]: DEBUG nova.scheduler.client.report [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.783455] env[62952]: INFO nova.compute.manager [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] [instance: 7c98a241-9297-417a-b412-5c279144572b] Took 1.04 seconds to deallocate network for instance. [ 672.922805] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.227026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.227026] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.228461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.082s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.230449] env[62952]: INFO nova.compute.claims [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.734787] env[62952]: DEBUG nova.compute.utils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.738526] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.738844] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 673.832498] env[62952]: DEBUG nova.policy [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '592ab470a2504c459a9f20f6e3c7a3bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71cf6551882a4c34ba621ee9b76efb8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.838554] env[62952]: INFO nova.scheduler.client.report [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Deleted allocations for instance 7c98a241-9297-417a-b412-5c279144572b [ 674.239704] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.351819] env[62952]: DEBUG oslo_concurrency.lockutils [None req-39d07b12-69d2-456e-9860-a187b056ac42 tempest-ServersTestFqdnHostnames-322977743 tempest-ServersTestFqdnHostnames-322977743-project-member] Lock "7c98a241-9297-417a-b412-5c279144572b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.852s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.379183] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Successfully created port: 06eef808-37db-449f-835e-c20e03e67bcb {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.387281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquiring lock "91660994-feca-40d8-8b75-d3c79110a19e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.387510] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Lock "91660994-feca-40d8-8b75-d3c79110a19e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.788805] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d9a1f6-f9db-4c9e-bfb1-84f267550d00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.797332] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb6e844-9b89-413a-b988-0f22421b6793 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.832174] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488a609b-07d3-4c3c-b0b5-71bca98327aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.836651] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6858488-cf12-4105-be3c-aa5564818815 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.852149] env[62952]: DEBUG nova.compute.provider_tree [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.853645] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.258508] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.293231] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.293537] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.293696] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.293874] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.294316] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.294550] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.294825] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.295084] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.295302] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.295472] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.295697] env[62952]: DEBUG nova.virt.hardware [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.296811] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d7ec75-3d9e-474d-9395-53b0a3d885a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.309117] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88026a4f-a38d-4afc-933d-8b07ad713513 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.358095] env[62952]: DEBUG nova.scheduler.client.report [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.385627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.478167] env[62952]: DEBUG nova.compute.manager [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] [instance: 288918a3-85be-416d-99fc-fd3450029749] Received event network-changed-06eef808-37db-449f-835e-c20e03e67bcb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.478430] env[62952]: DEBUG nova.compute.manager [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] [instance: 288918a3-85be-416d-99fc-fd3450029749] Refreshing instance network info cache due to event network-changed-06eef808-37db-449f-835e-c20e03e67bcb. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.478685] env[62952]: DEBUG oslo_concurrency.lockutils [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] Acquiring lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.478876] env[62952]: DEBUG oslo_concurrency.lockutils [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] Acquired lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.479221] env[62952]: DEBUG nova.network.neutron [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] [instance: 288918a3-85be-416d-99fc-fd3450029749] Refreshing network info cache for port 06eef808-37db-449f-835e-c20e03e67bcb {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 675.742878] env[62952]: ERROR nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 675.742878] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.742878] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.742878] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.742878] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.742878] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.742878] env[62952]: ERROR nova.compute.manager raise self.value [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.742878] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.742878] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.742878] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.743366] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.743366] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.743366] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 675.743366] env[62952]: ERROR nova.compute.manager [ 675.743366] env[62952]: Traceback (most recent call last): [ 675.743366] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.743366] env[62952]: listener.cb(fileno) [ 675.743366] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.743366] env[62952]: result = function(*args, **kwargs) [ 675.743366] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.743366] env[62952]: return func(*args, **kwargs) [ 675.743366] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.743366] env[62952]: raise e [ 675.743366] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.743366] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 675.743366] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.743366] env[62952]: created_port_ids = self._update_ports_for_instance( [ 675.743366] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.743366] env[62952]: with excutils.save_and_reraise_exception(): [ 675.743366] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.743366] env[62952]: self.force_reraise() [ 675.743366] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.743366] env[62952]: raise self.value [ 675.743366] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.743366] env[62952]: updated_port = self._update_port( [ 675.743366] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.743366] env[62952]: _ensure_no_port_binding_failure(port) [ 675.743366] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.743366] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.744070] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 675.744070] env[62952]: Removing descriptor: 15 [ 675.744070] env[62952]: ERROR nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] Traceback (most recent call last): [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] yield resources [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self.driver.spawn(context, instance, image_meta, [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.744070] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] vm_ref = self.build_virtual_machine(instance, [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] for vif in network_info: [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return self._sync_wrapper(fn, *args, **kwargs) [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self.wait() [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self[:] = self._gt.wait() [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return self._exit_event.wait() [ 675.744377] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] result = hub.switch() [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return self.greenlet.switch() [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] result = function(*args, **kwargs) [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return func(*args, **kwargs) [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] raise e [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] nwinfo = self.network_api.allocate_for_instance( [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.744703] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] created_port_ids = self._update_ports_for_instance( [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] with excutils.save_and_reraise_exception(): [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self.force_reraise() [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] raise self.value [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] updated_port = self._update_port( [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] _ensure_no_port_binding_failure(port) [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.745036] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] raise exception.PortBindingFailed(port_id=port['id']) [ 675.745345] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 675.745345] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] [ 675.745345] env[62952]: INFO nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Terminating instance [ 675.747147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.868162] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.868162] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.868450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.956s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.999489] env[62952]: DEBUG nova.network.neutron [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.094855] env[62952]: DEBUG nova.network.neutron [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] [instance: 288918a3-85be-416d-99fc-fd3450029749] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.373458] env[62952]: DEBUG nova.compute.utils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 676.377884] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 676.378096] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 676.489234] env[62952]: DEBUG nova.policy [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ae217ec66fe45beb1381449073980d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23e42f4ddc5a4d96b3682876928f1dfc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.599082] env[62952]: DEBUG oslo_concurrency.lockutils [req-1ab07077-f0c3-467d-aee2-2197a7b5ad67 req-dc8116eb-d8f8-4d5a-9154-0341e95aa926 service nova] Releasing lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.599920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquired lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.600583] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.714526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquiring lock "c7771b47-cbe8-4696-9a21-9204501c8877" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.714708] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Lock "c7771b47-cbe8-4696-9a21-9204501c8877" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.872698] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8ee18a-541a-478f-a45c-e20d7d8d1d88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.881444] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.885501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d4a1c6-8a1a-4b6b-835d-9f7220ca6427 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.915345] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc72d058-270d-4682-bf38-9d4f4e80289e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.924087] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f05579d-5b83-4947-a733-bb2fc5484936 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.939649] env[62952]: DEBUG nova.compute.provider_tree [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.136358] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.285149] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Successfully created port: 49503947-c159-440f-ac60-05712e35899b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 677.357274] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.442852] env[62952]: DEBUG nova.scheduler.client.report [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.545737] env[62952]: DEBUG nova.compute.manager [req-9f22b228-2fb5-441e-ae5f-d17adee4f685 req-a6986dc8-8878-4799-b8c7-9433cf65e7bc service nova] [instance: 288918a3-85be-416d-99fc-fd3450029749] Received event network-vif-deleted-06eef808-37db-449f-835e-c20e03e67bcb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.864022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Releasing lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.864503] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.864503] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.864805] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a27b912b-b1bd-41d3-9497-a439153a64e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.873785] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f46e916-9403-43c2-afca-1be91592bcae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.895595] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 288918a3-85be-416d-99fc-fd3450029749 could not be found. [ 677.895820] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 677.896032] env[62952]: INFO nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Took 0.03 seconds to destroy the instance on the hypervisor. [ 677.896308] env[62952]: DEBUG oslo.service.loopingcall [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.897354] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.899296] env[62952]: DEBUG nova.compute.manager [-] [instance: 288918a3-85be-416d-99fc-fd3450029749] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.899399] env[62952]: DEBUG nova.network.neutron [-] [instance: 288918a3-85be-416d-99fc-fd3450029749] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.919952] env[62952]: DEBUG nova.network.neutron [-] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.929194] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.929457] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.929616] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.929849] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.930096] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.930096] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.930295] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.930471] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.930604] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.930758] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.930998] env[62952]: DEBUG nova.virt.hardware [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.931775] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae633b5-cf96-42df-9d93-e674e7971b99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.940559] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5977625-2c79-401e-b796-ecb23ebacbb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.954850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.086s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.955192] env[62952]: ERROR nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Traceback (most recent call last): [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self.driver.spawn(context, instance, image_meta, [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] vm_ref = self.build_virtual_machine(instance, [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.955192] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] for vif in network_info: [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return self._sync_wrapper(fn, *args, **kwargs) [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self.wait() [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self[:] = self._gt.wait() [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return self._exit_event.wait() [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] result = hub.switch() [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.955661] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return self.greenlet.switch() [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] result = function(*args, **kwargs) [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] return func(*args, **kwargs) [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] raise e [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] nwinfo = self.network_api.allocate_for_instance( [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] created_port_ids = self._update_ports_for_instance( [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] with excutils.save_and_reraise_exception(): [ 677.956222] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] self.force_reraise() [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] raise self.value [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] updated_port = self._update_port( [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] _ensure_no_port_binding_failure(port) [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] raise exception.PortBindingFailed(port_id=port['id']) [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] nova.exception.PortBindingFailed: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. [ 677.956802] env[62952]: ERROR nova.compute.manager [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] [ 677.959460] env[62952]: DEBUG nova.compute.utils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.961155] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.825s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.962618] env[62952]: INFO nova.compute.claims [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.965210] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Build of instance 542169bf-71e2-47f8-a311-8a3dca4ffee7 was re-scheduled: Binding failed for port 85839fde-aea9-4fe2-8bad-808ff10fcc08, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.965660] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.965883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquiring lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.966050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Acquired lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.966215] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.344022] env[62952]: ERROR nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 678.344022] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.344022] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.344022] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.344022] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.344022] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.344022] env[62952]: ERROR nova.compute.manager raise self.value [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.344022] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 678.344022] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.344022] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 678.344489] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.344489] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 678.344489] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 678.344489] env[62952]: ERROR nova.compute.manager [ 678.344489] env[62952]: Traceback (most recent call last): [ 678.344489] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 678.344489] env[62952]: listener.cb(fileno) [ 678.344489] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.344489] env[62952]: result = function(*args, **kwargs) [ 678.344489] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.344489] env[62952]: return func(*args, **kwargs) [ 678.344489] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.344489] env[62952]: raise e [ 678.344489] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.344489] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 678.344489] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.344489] env[62952]: created_port_ids = self._update_ports_for_instance( [ 678.344489] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.344489] env[62952]: with excutils.save_and_reraise_exception(): [ 678.344489] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.344489] env[62952]: self.force_reraise() [ 678.344489] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.344489] env[62952]: raise self.value [ 678.344489] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.344489] env[62952]: updated_port = self._update_port( [ 678.344489] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.344489] env[62952]: _ensure_no_port_binding_failure(port) [ 678.344489] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.344489] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 678.345229] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 678.345229] env[62952]: Removing descriptor: 15 [ 678.345229] env[62952]: ERROR nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] Traceback (most recent call last): [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] yield resources [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self.driver.spawn(context, instance, image_meta, [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.345229] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] vm_ref = self.build_virtual_machine(instance, [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] for vif in network_info: [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return self._sync_wrapper(fn, *args, **kwargs) [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self.wait() [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self[:] = self._gt.wait() [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return self._exit_event.wait() [ 678.345573] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] result = hub.switch() [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return self.greenlet.switch() [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] result = function(*args, **kwargs) [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return func(*args, **kwargs) [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] raise e [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] nwinfo = self.network_api.allocate_for_instance( [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.345845] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] created_port_ids = self._update_ports_for_instance( [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] with excutils.save_and_reraise_exception(): [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self.force_reraise() [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] raise self.value [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] updated_port = self._update_port( [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] _ensure_no_port_binding_failure(port) [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.346173] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] raise exception.PortBindingFailed(port_id=port['id']) [ 678.346484] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 678.346484] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] [ 678.346484] env[62952]: INFO nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Terminating instance [ 678.346484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.346484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquired lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.346625] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.423453] env[62952]: DEBUG nova.network.neutron [-] [instance: 288918a3-85be-416d-99fc-fd3450029749] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.487313] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.547297] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.865155] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.925961] env[62952]: INFO nova.compute.manager [-] [instance: 288918a3-85be-416d-99fc-fd3450029749] Took 1.03 seconds to deallocate network for instance. [ 678.928421] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.929610] env[62952]: DEBUG nova.compute.claims [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.929777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.050605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Releasing lock "refresh_cache-542169bf-71e2-47f8-a311-8a3dca4ffee7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.050919] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 679.051174] env[62952]: DEBUG nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.051654] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.069798] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.391961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b061a2a-73cc-48d0-a7fe-f1ae3c6122b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.401017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532a2368-c941-4f2d-b2f0-6fb7676b726c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.429937] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f27fc4f-2998-48c2-973f-cd69283e201a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.432664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Releasing lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.433075] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.433270] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.433518] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd3be4e2-c738-4cfc-bad8-c9237d8e5cec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.440652] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfd9860-830c-4105-9452-53deee5bf128 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.446561] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d674131-a8a7-4ece-8266-b467381903a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.466724] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d8653006-b429-44b5-b978-ee00a279e50e could not be found. [ 679.466940] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.467131] env[62952]: INFO nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 679.467365] env[62952]: DEBUG oslo.service.loopingcall [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.474486] env[62952]: DEBUG nova.compute.manager [-] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.474586] env[62952]: DEBUG nova.network.neutron [-] [instance: d8653006-b429-44b5-b978-ee00a279e50e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.476341] env[62952]: DEBUG nova.compute.provider_tree [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.491351] env[62952]: DEBUG nova.network.neutron [-] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.572119] env[62952]: DEBUG nova.network.neutron [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.608049] env[62952]: DEBUG nova.compute.manager [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Received event network-changed-49503947-c159-440f-ac60-05712e35899b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.608348] env[62952]: DEBUG nova.compute.manager [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Refreshing instance network info cache due to event network-changed-49503947-c159-440f-ac60-05712e35899b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 679.608568] env[62952]: DEBUG oslo_concurrency.lockutils [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] Acquiring lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.608711] env[62952]: DEBUG oslo_concurrency.lockutils [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] Acquired lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.608870] env[62952]: DEBUG nova.network.neutron [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Refreshing network info cache for port 49503947-c159-440f-ac60-05712e35899b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 679.979993] env[62952]: DEBUG nova.scheduler.client.report [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.993860] env[62952]: DEBUG nova.network.neutron [-] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.075049] env[62952]: INFO nova.compute.manager [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] [instance: 542169bf-71e2-47f8-a311-8a3dca4ffee7] Took 1.02 seconds to deallocate network for instance. [ 680.129726] env[62952]: DEBUG nova.network.neutron [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.174601] env[62952]: DEBUG nova.network.neutron [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.487096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.487096] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.488150] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.276s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.493964] env[62952]: INFO nova.compute.claims [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.498320] env[62952]: INFO nova.compute.manager [-] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Took 1.02 seconds to deallocate network for instance. [ 680.500575] env[62952]: DEBUG nova.compute.claims [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.500851] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.677328] env[62952]: DEBUG oslo_concurrency.lockutils [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] Releasing lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.677611] env[62952]: DEBUG nova.compute.manager [req-578dac63-5877-43f4-8ca8-45a1f60baae3 req-7183f6ba-4ca6-4bac-af96-3a550f325489 service nova] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Received event network-vif-deleted-49503947-c159-440f-ac60-05712e35899b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.001940] env[62952]: DEBUG nova.compute.utils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.003633] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 681.005957] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 681.048768] env[62952]: DEBUG nova.policy [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b019273ad7e414c96457e96391138bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c57a43be3c844332b7211f8200cc48db', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 681.108047] env[62952]: INFO nova.scheduler.client.report [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Deleted allocations for instance 542169bf-71e2-47f8-a311-8a3dca4ffee7 [ 681.326852] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Successfully created port: c2cb65ec-f4c4-43ee-abfe-11c6c590e929 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.505074] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 681.618281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb76a1c-021f-463f-b88a-3de8642835cf tempest-ServerDiagnosticsNegativeTest-777876466 tempest-ServerDiagnosticsNegativeTest-777876466-project-member] Lock "542169bf-71e2-47f8-a311-8a3dca4ffee7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.425s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.926181] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98337867-0d3c-4ce9-9b99-5ecd8297fc8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.938897] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613e74dc-8644-404d-83eb-4247127a7aca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.976734] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280cfb47-8f46-45a3-88e3-f521cdf71a7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.987886] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632ce230-3526-4a96-bbc5-931112e7a59d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.002133] env[62952]: DEBUG nova.compute.provider_tree [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.123252] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.165956] env[62952]: DEBUG nova.compute.manager [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Received event network-changed-c2cb65ec-f4c4-43ee-abfe-11c6c590e929 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.166172] env[62952]: DEBUG nova.compute.manager [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Refreshing instance network info cache due to event network-changed-c2cb65ec-f4c4-43ee-abfe-11c6c590e929. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 682.166384] env[62952]: DEBUG oslo_concurrency.lockutils [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] Acquiring lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.166528] env[62952]: DEBUG oslo_concurrency.lockutils [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] Acquired lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.166686] env[62952]: DEBUG nova.network.neutron [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Refreshing network info cache for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 682.347851] env[62952]: ERROR nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 682.347851] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.347851] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.347851] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.347851] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.347851] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.347851] env[62952]: ERROR nova.compute.manager raise self.value [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.347851] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 682.347851] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.347851] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 682.348394] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.348394] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 682.348394] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 682.348394] env[62952]: ERROR nova.compute.manager [ 682.348394] env[62952]: Traceback (most recent call last): [ 682.348394] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 682.348394] env[62952]: listener.cb(fileno) [ 682.348394] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.348394] env[62952]: result = function(*args, **kwargs) [ 682.348394] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.348394] env[62952]: return func(*args, **kwargs) [ 682.348394] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.348394] env[62952]: raise e [ 682.348394] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.348394] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 682.348394] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.348394] env[62952]: created_port_ids = self._update_ports_for_instance( [ 682.348394] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.348394] env[62952]: with excutils.save_and_reraise_exception(): [ 682.348394] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.348394] env[62952]: self.force_reraise() [ 682.348394] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.348394] env[62952]: raise self.value [ 682.348394] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.348394] env[62952]: updated_port = self._update_port( [ 682.348394] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.348394] env[62952]: _ensure_no_port_binding_failure(port) [ 682.348394] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.348394] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 682.349043] env[62952]: nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 682.349043] env[62952]: Removing descriptor: 15 [ 682.505630] env[62952]: DEBUG nova.scheduler.client.report [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.520529] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.549252] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.549530] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.549694] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.549878] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.550089] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.550255] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.550450] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.550610] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.550773] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.550931] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.551111] env[62952]: DEBUG nova.virt.hardware [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.552077] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b1af31-2452-40db-ab80-fb7ce652c9cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.561853] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54892fce-071b-4add-af48-88e3e3732011 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.574577] env[62952]: ERROR nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Traceback (most recent call last): [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] yield resources [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self.driver.spawn(context, instance, image_meta, [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] vm_ref = self.build_virtual_machine(instance, [ 682.574577] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] for vif in network_info: [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] return self._sync_wrapper(fn, *args, **kwargs) [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self.wait() [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self[:] = self._gt.wait() [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] return self._exit_event.wait() [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.574887] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] current.throw(*self._exc) [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] result = function(*args, **kwargs) [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] return func(*args, **kwargs) [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] raise e [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] nwinfo = self.network_api.allocate_for_instance( [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] created_port_ids = self._update_ports_for_instance( [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] with excutils.save_and_reraise_exception(): [ 682.575217] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self.force_reraise() [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] raise self.value [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] updated_port = self._update_port( [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] _ensure_no_port_binding_failure(port) [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] raise exception.PortBindingFailed(port_id=port['id']) [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 682.575529] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] [ 682.575529] env[62952]: INFO nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Terminating instance [ 682.576888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquiring lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.647923] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.685232] env[62952]: DEBUG nova.network.neutron [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.779983] env[62952]: DEBUG nova.network.neutron [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.011415] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.011934] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 683.014892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.651s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.016390] env[62952]: INFO nova.compute.claims [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.761133] env[62952]: DEBUG oslo_concurrency.lockutils [req-df3a2c23-d7a2-4d5f-bf6b-8a132d3cdd89 req-a33e1d3d-1296-4328-81f4-e2e1037e8fc6 service nova] Releasing lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.764629] env[62952]: DEBUG nova.compute.utils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.765948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquired lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.769034] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.769034] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 683.769034] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 683.856605] env[62952]: DEBUG nova.policy [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2421d5410e274cdda439e0851d751db2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a873aeba5284b0cb64bdf19b28e38d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.248296] env[62952]: DEBUG nova.compute.manager [req-c4b8bb04-6854-47fb-a3cd-b8f0cecfec47 req-cbe73157-dd04-4f9f-8228-431ec8dfc51e service nova] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Received event network-vif-deleted-c2cb65ec-f4c4-43ee-abfe-11c6c590e929 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.258240] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Successfully created port: f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 684.271601] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 684.292544] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.428828] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.687149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52664e74-4b8b-4794-977d-fa38e34ec1ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.697468] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe4653b-9717-4c90-9c3c-8d0c29a9960b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.730124] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9243b2-0831-4886-8419-6729250b7f8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.738801] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bca4ab2-ee84-444a-9e56-ddf37f4507c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.754631] env[62952]: DEBUG nova.compute.provider_tree [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.931495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Releasing lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.931972] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 684.932186] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.932572] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6a3ae25-52da-4007-8afb-2e2940c701ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.941591] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649efdfa-9c95-4b00-873a-1cdb18514d3a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.965969] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8db47d06-230f-4c5e-8bd6-9d5b543455b6 could not be found. [ 684.966278] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.966548] env[62952]: INFO nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 684.966819] env[62952]: DEBUG oslo.service.loopingcall [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.967055] env[62952]: DEBUG nova.compute.manager [-] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.967150] env[62952]: DEBUG nova.network.neutron [-] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.994216] env[62952]: DEBUG nova.network.neutron [-] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.258339] env[62952]: DEBUG nova.scheduler.client.report [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.281659] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 685.309651] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 685.309900] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 685.310077] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.310354] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 685.310427] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.310572] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 685.310761] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 685.310941] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 685.311281] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 685.311468] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 685.311645] env[62952]: DEBUG nova.virt.hardware [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 685.312482] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc4708e-9a89-47be-9035-94ef6aec585e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.320673] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9165513-03aa-4e31-ba14-53e1252a68f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.497649] env[62952]: DEBUG nova.network.neutron [-] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.767883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.753s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.768455] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.771071] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.070s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.773667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.773667] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 685.773667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.370s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.780290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cea6f52-451c-4b65-b18b-3a5d65775c38 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.796618] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40237a9b-e8f7-4ff6-b8a2-6d307c200ad4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.813628] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab04ac7-3f51-4c8b-94f8-1cd64d8f1611 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.821048] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d944ea8-40c1-427f-b5bc-7e79614379f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.855742] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181500MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 685.855909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.000039] env[62952]: INFO nova.compute.manager [-] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Took 1.03 seconds to deallocate network for instance. [ 686.004905] env[62952]: DEBUG nova.compute.claims [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 686.004905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.219147] env[62952]: ERROR nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 686.219147] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.219147] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.219147] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.219147] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.219147] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.219147] env[62952]: ERROR nova.compute.manager raise self.value [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.219147] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.219147] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.219147] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.219551] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.219551] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.219551] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 686.219551] env[62952]: ERROR nova.compute.manager [ 686.219551] env[62952]: Traceback (most recent call last): [ 686.219551] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.219551] env[62952]: listener.cb(fileno) [ 686.219551] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.219551] env[62952]: result = function(*args, **kwargs) [ 686.219551] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.219551] env[62952]: return func(*args, **kwargs) [ 686.219551] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.219551] env[62952]: raise e [ 686.219551] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.219551] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 686.219551] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.219551] env[62952]: created_port_ids = self._update_ports_for_instance( [ 686.219551] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.219551] env[62952]: with excutils.save_and_reraise_exception(): [ 686.219551] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.219551] env[62952]: self.force_reraise() [ 686.219551] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.219551] env[62952]: raise self.value [ 686.219551] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.219551] env[62952]: updated_port = self._update_port( [ 686.219551] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.219551] env[62952]: _ensure_no_port_binding_failure(port) [ 686.219551] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.219551] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.220512] env[62952]: nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 686.220512] env[62952]: Removing descriptor: 15 [ 686.220512] env[62952]: ERROR nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Traceback (most recent call last): [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] yield resources [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self.driver.spawn(context, instance, image_meta, [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.220512] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] vm_ref = self.build_virtual_machine(instance, [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] for vif in network_info: [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return self._sync_wrapper(fn, *args, **kwargs) [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self.wait() [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self[:] = self._gt.wait() [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return self._exit_event.wait() [ 686.220824] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] result = hub.switch() [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return self.greenlet.switch() [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] result = function(*args, **kwargs) [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return func(*args, **kwargs) [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] raise e [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] nwinfo = self.network_api.allocate_for_instance( [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.221225] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] created_port_ids = self._update_ports_for_instance( [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] with excutils.save_and_reraise_exception(): [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self.force_reraise() [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] raise self.value [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] updated_port = self._update_port( [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] _ensure_no_port_binding_failure(port) [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.221554] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] raise exception.PortBindingFailed(port_id=port['id']) [ 686.221856] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 686.221856] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] [ 686.221856] env[62952]: INFO nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Terminating instance [ 686.222474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquiring lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.222635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquired lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.222801] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.274228] env[62952]: DEBUG nova.compute.manager [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Received event network-changed-f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.274228] env[62952]: DEBUG nova.compute.manager [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Refreshing instance network info cache due to event network-changed-f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 686.274228] env[62952]: DEBUG oslo_concurrency.lockutils [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] Acquiring lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.287993] env[62952]: DEBUG nova.compute.utils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.290308] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 686.290499] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.377560] env[62952]: DEBUG nova.policy [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84a933f0beb740bb96a0ffab4c44568b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c6d16afcdb245ff97759f9b6ed1ed43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 686.713879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17a8a8a-0ddf-4b73-9818-cc65e8ddb6e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.724823] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856718b6-29fb-414b-893c-bc8253d892d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.759642] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Successfully created port: d9c5eae3-096a-4b17-a34b-0874b5bd8c36 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.761991] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.764972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfec55a-8af8-4c52-a987-28c2f6ec3be5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.772805] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4896706f-4b50-4a3a-9e95-cb71e2e4577a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.786879] env[62952]: DEBUG nova.compute.provider_tree [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.793852] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.854799] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.292810] env[62952]: DEBUG nova.scheduler.client.report [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.352261] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "4be2f0a4-5f57-4a80-9450-6d61d344799e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.352461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "4be2f0a4-5f57-4a80-9450-6d61d344799e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.356490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Releasing lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.356874] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.357078] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.357370] env[62952]: DEBUG oslo_concurrency.lockutils [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] Acquired lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.357540] env[62952]: DEBUG nova.network.neutron [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Refreshing network info cache for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.358531] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43a55fad-9735-4385-a27b-5e4afb43c7ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.369289] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeee56f-845f-4703-8aac-f4aa118fc7b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.395034] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f54e2a1f-987b-40ac-8197-d6e4ad3ba18f could not be found. [ 687.395317] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.395547] env[62952]: INFO nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 687.395835] env[62952]: DEBUG oslo.service.loopingcall [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.396353] env[62952]: DEBUG nova.compute.manager [-] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.396557] env[62952]: DEBUG nova.network.neutron [-] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.418173] env[62952]: DEBUG nova.network.neutron [-] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.802676] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.031s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.805022] env[62952]: ERROR nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Traceback (most recent call last): [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self.driver.spawn(context, instance, image_meta, [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] vm_ref = self.build_virtual_machine(instance, [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.805022] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] for vif in network_info: [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return self._sync_wrapper(fn, *args, **kwargs) [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self.wait() [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self[:] = self._gt.wait() [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return self._exit_event.wait() [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] result = hub.switch() [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.805399] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return self.greenlet.switch() [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] result = function(*args, **kwargs) [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] return func(*args, **kwargs) [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] raise e [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] nwinfo = self.network_api.allocate_for_instance( [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] created_port_ids = self._update_ports_for_instance( [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] with excutils.save_and_reraise_exception(): [ 687.805755] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] self.force_reraise() [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] raise self.value [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] updated_port = self._update_port( [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] _ensure_no_port_binding_failure(port) [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] raise exception.PortBindingFailed(port_id=port['id']) [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] nova.exception.PortBindingFailed: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. [ 687.806199] env[62952]: ERROR nova.compute.manager [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] [ 687.806490] env[62952]: DEBUG nova.compute.utils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.810177] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.811597] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Build of instance 12408f00-ed1d-4b96-bc0c-59779d0eda74 was re-scheduled: Binding failed for port ba13b6c5-0256-4df0-b394-7d07b2970c0d, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.812172] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.814018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquiring lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.814018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Acquired lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.814018] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.814018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.677s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.816086] env[62952]: INFO nova.compute.claims [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.843897] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:34:28Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='82709d86-89cc-457d-a896-01beea9a7fe0',id=25,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1483123089',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.844162] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.844320] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.844499] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.844642] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.844788] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.844992] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.845166] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.845332] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.845492] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.845664] env[62952]: DEBUG nova.virt.hardware [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.846538] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85896f2b-40f7-4475-b55a-44443a62795a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.851522] env[62952]: ERROR nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 687.851522] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.851522] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.851522] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.851522] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.851522] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.851522] env[62952]: ERROR nova.compute.manager raise self.value [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.851522] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.851522] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.851522] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.852071] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.852071] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.852071] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 687.852071] env[62952]: ERROR nova.compute.manager [ 687.852071] env[62952]: Traceback (most recent call last): [ 687.852071] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.852071] env[62952]: listener.cb(fileno) [ 687.852071] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.852071] env[62952]: result = function(*args, **kwargs) [ 687.852071] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.852071] env[62952]: return func(*args, **kwargs) [ 687.852071] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.852071] env[62952]: raise e [ 687.852071] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.852071] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 687.852071] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.852071] env[62952]: created_port_ids = self._update_ports_for_instance( [ 687.852071] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.852071] env[62952]: with excutils.save_and_reraise_exception(): [ 687.852071] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.852071] env[62952]: self.force_reraise() [ 687.852071] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.852071] env[62952]: raise self.value [ 687.852071] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.852071] env[62952]: updated_port = self._update_port( [ 687.852071] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.852071] env[62952]: _ensure_no_port_binding_failure(port) [ 687.852071] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.852071] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.852825] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 687.852825] env[62952]: Removing descriptor: 15 [ 687.857812] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0287c1b1-3c3f-41ba-9128-d778adfba996 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.874539] env[62952]: ERROR nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Traceback (most recent call last): [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] yield resources [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self.driver.spawn(context, instance, image_meta, [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] vm_ref = self.build_virtual_machine(instance, [ 687.874539] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] for vif in network_info: [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] return self._sync_wrapper(fn, *args, **kwargs) [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self.wait() [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self[:] = self._gt.wait() [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] return self._exit_event.wait() [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.874889] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] current.throw(*self._exc) [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] result = function(*args, **kwargs) [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] return func(*args, **kwargs) [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] raise e [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] nwinfo = self.network_api.allocate_for_instance( [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] created_port_ids = self._update_ports_for_instance( [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] with excutils.save_and_reraise_exception(): [ 687.875249] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self.force_reraise() [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] raise self.value [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] updated_port = self._update_port( [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] _ensure_no_port_binding_failure(port) [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] raise exception.PortBindingFailed(port_id=port['id']) [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 687.875602] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] [ 687.875602] env[62952]: INFO nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Terminating instance [ 687.876642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.876827] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquired lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.876998] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.878583] env[62952]: DEBUG nova.network.neutron [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.919929] env[62952]: DEBUG nova.network.neutron [-] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.923813] env[62952]: DEBUG nova.network.neutron [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.304518] env[62952]: DEBUG nova.compute.manager [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Received event network-changed-d9c5eae3-096a-4b17-a34b-0874b5bd8c36 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.304811] env[62952]: DEBUG nova.compute.manager [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Refreshing instance network info cache due to event network-changed-d9c5eae3-096a-4b17-a34b-0874b5bd8c36. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 688.304947] env[62952]: DEBUG oslo_concurrency.lockutils [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] Acquiring lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.337621] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.411879] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.422008] env[62952]: INFO nova.compute.manager [-] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Took 1.03 seconds to deallocate network for instance. [ 688.424127] env[62952]: DEBUG nova.compute.claims [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.424309] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.427485] env[62952]: DEBUG oslo_concurrency.lockutils [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] Releasing lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.427699] env[62952]: DEBUG nova.compute.manager [req-6a936d3d-7576-401c-95cf-860036982ef3 req-1cfcd4fc-9362-4b31-b044-8362c574b660 service nova] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Received event network-vif-deleted-f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.442855] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.481914] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.945489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Releasing lock "refresh_cache-12408f00-ed1d-4b96-bc0c-59779d0eda74" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.945729] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 688.945907] env[62952]: DEBUG nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.946087] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.962962] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.984506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Releasing lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.984972] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 688.985107] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 688.985402] env[62952]: DEBUG oslo_concurrency.lockutils [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] Acquired lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.985571] env[62952]: DEBUG nova.network.neutron [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Refreshing network info cache for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.986735] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbd39671-ab3e-494a-8ad0-27428bc1a66f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.998304] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be3a12f-b87d-46f1-bcb0-7dbab1f409b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.022877] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 40b6590d-44e1-4d02-b1fa-46c1e6c861ea could not be found. [ 689.023108] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.023297] env[62952]: INFO nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 689.023533] env[62952]: DEBUG oslo.service.loopingcall [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.025759] env[62952]: DEBUG nova.compute.manager [-] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.025817] env[62952]: DEBUG nova.network.neutron [-] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.043255] env[62952]: DEBUG nova.network.neutron [-] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.179349] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42b9f06-3088-4cd1-a9c9-443cbc9766af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.186578] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7ff616-120f-4dfd-97a2-60c2827c54ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.215530] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddaaefd-8235-43a8-9481-f2c99ada79fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.222724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb8d869-1873-4a00-97f2-1a40b1d72d69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.235446] env[62952]: DEBUG nova.compute.provider_tree [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.466027] env[62952]: DEBUG nova.network.neutron [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.502681] env[62952]: DEBUG nova.network.neutron [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.546593] env[62952]: DEBUG nova.network.neutron [-] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.560348] env[62952]: DEBUG nova.network.neutron [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.738590] env[62952]: DEBUG nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.969323] env[62952]: INFO nova.compute.manager [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] [instance: 12408f00-ed1d-4b96-bc0c-59779d0eda74] Took 1.02 seconds to deallocate network for instance. [ 690.049116] env[62952]: INFO nova.compute.manager [-] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Took 1.02 seconds to deallocate network for instance. [ 690.051314] env[62952]: DEBUG nova.compute.claims [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.051484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.062040] env[62952]: DEBUG oslo_concurrency.lockutils [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] Releasing lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.062253] env[62952]: DEBUG nova.compute.manager [req-595b0369-25ec-4b94-8497-dac0fb42ea35 req-67c0e5b4-5c8e-4813-89d1-c874d28fe8ba service nova] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Received event network-vif-deleted-d9c5eae3-096a-4b17-a34b-0874b5bd8c36 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.243362] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.243876] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.246506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.324s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.247901] env[62952]: INFO nova.compute.claims [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.752299] env[62952]: DEBUG nova.compute.utils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.753760] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.753933] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 690.796973] env[62952]: DEBUG nova.policy [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4775d366579c453080fe8686fc9f9adb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d446121953c4ca99acd8bd14b9168e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.000942] env[62952]: INFO nova.scheduler.client.report [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Deleted allocations for instance 12408f00-ed1d-4b96-bc0c-59779d0eda74 [ 691.126156] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Successfully created port: 6789cad3-4447-4caf-a77a-4151a1974032 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.258723] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.509823] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dfe3e43f-b26f-4fcd-96e1-fd2d45c84bda tempest-ServersAdminTestJSON-102436230 tempest-ServersAdminTestJSON-102436230-project-member] Lock "12408f00-ed1d-4b96-bc0c-59779d0eda74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.277s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.653668] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f8d923-b017-44be-b45c-dac5b72c8b39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.661575] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e8b422-3c5b-42e3-989a-52dce3c5e067 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.693348] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afedc55-aadf-4e78-ade4-07d2f1835cec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.701450] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828651dd-5da7-4169-b8f2-6038aeb89b17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.715881] env[62952]: DEBUG nova.compute.provider_tree [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.012807] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.219279] env[62952]: DEBUG nova.scheduler.client.report [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.256615] env[62952]: DEBUG nova.compute.manager [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Received event network-changed-6789cad3-4447-4caf-a77a-4151a1974032 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.256817] env[62952]: DEBUG nova.compute.manager [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Refreshing instance network info cache due to event network-changed-6789cad3-4447-4caf-a77a-4151a1974032. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.257044] env[62952]: DEBUG oslo_concurrency.lockutils [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] Acquiring lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.257191] env[62952]: DEBUG oslo_concurrency.lockutils [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] Acquired lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.257349] env[62952]: DEBUG nova.network.neutron [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Refreshing network info cache for port 6789cad3-4447-4caf-a77a-4151a1974032 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 692.270146] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.301188] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.301450] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.301607] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.301785] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.301929] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.302132] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.302356] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.302513] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.302675] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.302833] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.303007] env[62952]: DEBUG nova.virt.hardware [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.304083] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af62c4f4-9b74-4b6a-88d7-5f9de9ff99df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.312635] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf551d8-44a0-46d8-97f3-f051fe5b09d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.541405] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.561879] env[62952]: ERROR nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 692.561879] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.561879] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.561879] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.561879] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.561879] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.561879] env[62952]: ERROR nova.compute.manager raise self.value [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.561879] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.561879] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.561879] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.562338] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.562338] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.562338] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 692.562338] env[62952]: ERROR nova.compute.manager [ 692.562338] env[62952]: Traceback (most recent call last): [ 692.562338] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.562338] env[62952]: listener.cb(fileno) [ 692.562338] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.562338] env[62952]: result = function(*args, **kwargs) [ 692.562338] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.562338] env[62952]: return func(*args, **kwargs) [ 692.562338] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.562338] env[62952]: raise e [ 692.562338] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.562338] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 692.562338] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.562338] env[62952]: created_port_ids = self._update_ports_for_instance( [ 692.562338] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.562338] env[62952]: with excutils.save_and_reraise_exception(): [ 692.562338] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.562338] env[62952]: self.force_reraise() [ 692.562338] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.562338] env[62952]: raise self.value [ 692.562338] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.562338] env[62952]: updated_port = self._update_port( [ 692.562338] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.562338] env[62952]: _ensure_no_port_binding_failure(port) [ 692.562338] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.562338] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.563109] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 692.563109] env[62952]: Removing descriptor: 19 [ 692.563109] env[62952]: ERROR nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Traceback (most recent call last): [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] yield resources [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self.driver.spawn(context, instance, image_meta, [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.563109] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] vm_ref = self.build_virtual_machine(instance, [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] for vif in network_info: [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return self._sync_wrapper(fn, *args, **kwargs) [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self.wait() [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self[:] = self._gt.wait() [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return self._exit_event.wait() [ 692.563441] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] result = hub.switch() [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return self.greenlet.switch() [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] result = function(*args, **kwargs) [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return func(*args, **kwargs) [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] raise e [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] nwinfo = self.network_api.allocate_for_instance( [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.563790] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] created_port_ids = self._update_ports_for_instance( [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] with excutils.save_and_reraise_exception(): [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self.force_reraise() [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] raise self.value [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] updated_port = self._update_port( [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] _ensure_no_port_binding_failure(port) [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.564160] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] raise exception.PortBindingFailed(port_id=port['id']) [ 692.564503] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 692.564503] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] [ 692.564503] env[62952]: INFO nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Terminating instance [ 692.569532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquiring lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.725810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.726387] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.729322] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.344s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.730788] env[62952]: INFO nova.compute.claims [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.782485] env[62952]: DEBUG nova.network.neutron [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.960931] env[62952]: DEBUG nova.network.neutron [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.245082] env[62952]: DEBUG nova.compute.utils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.248168] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.248168] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.309904] env[62952]: DEBUG nova.policy [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '133f56aa710045a5994c2effb34e7387', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd846730e0a3c4d748590225789bfa1d0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.465463] env[62952]: DEBUG oslo_concurrency.lockutils [req-5531635c-88d0-45f1-aefe-c98083fdf2de req-145fe63e-49e2-482b-8dc2-63e0ba6ce222 service nova] Releasing lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.465463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquired lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.465463] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.601269] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Successfully created port: 003856df-04d1-443c-968c-8e4f6c633926 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.752690] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.993160] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.132195] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.180951] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25a4b93-d46f-45a7-9d1e-23988f84f67a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.190029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85462f4e-f453-4209-a760-22ab77b762f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.222042] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693ec7db-d355-4d59-8053-bbed9f7f880e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.229984] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79cc44c-9fa8-459c-aa8c-80f2e8d67516 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.244042] env[62952]: DEBUG nova.compute.provider_tree [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.282391] env[62952]: DEBUG nova.compute.manager [req-82a28031-f7e2-4316-8d60-5071f48a62e6 req-21b2a0da-a968-481d-9fd8-e2fc5be26b46 service nova] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Received event network-vif-deleted-6789cad3-4447-4caf-a77a-4151a1974032 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.638397] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Releasing lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.638833] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.639114] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.639430] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abdaade6-4efd-41ac-b944-c6ff63637dba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.649880] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2469db-ad29-4d6d-956f-7eb355f0f1b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.671804] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cb35ba14-6a78-4381-9c93-f2fa285468d7 could not be found. [ 694.672018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.672206] env[62952]: INFO nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 694.672456] env[62952]: DEBUG oslo.service.loopingcall [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.672732] env[62952]: DEBUG nova.compute.manager [-] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.672956] env[62952]: DEBUG nova.network.neutron [-] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.690585] env[62952]: DEBUG nova.network.neutron [-] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.736914] env[62952]: ERROR nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 694.736914] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.736914] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.736914] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.736914] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.736914] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.736914] env[62952]: ERROR nova.compute.manager raise self.value [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.736914] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.736914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.736914] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.737431] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.737431] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.737431] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 694.737431] env[62952]: ERROR nova.compute.manager [ 694.737431] env[62952]: Traceback (most recent call last): [ 694.737431] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.737431] env[62952]: listener.cb(fileno) [ 694.737431] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.737431] env[62952]: result = function(*args, **kwargs) [ 694.737431] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.737431] env[62952]: return func(*args, **kwargs) [ 694.737431] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.737431] env[62952]: raise e [ 694.737431] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.737431] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 694.737431] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.737431] env[62952]: created_port_ids = self._update_ports_for_instance( [ 694.737431] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.737431] env[62952]: with excutils.save_and_reraise_exception(): [ 694.737431] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.737431] env[62952]: self.force_reraise() [ 694.737431] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.737431] env[62952]: raise self.value [ 694.737431] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.737431] env[62952]: updated_port = self._update_port( [ 694.737431] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.737431] env[62952]: _ensure_no_port_binding_failure(port) [ 694.737431] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.737431] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.738176] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 694.738176] env[62952]: Removing descriptor: 19 [ 694.747606] env[62952]: DEBUG nova.scheduler.client.report [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.762432] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.790456] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.790711] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.790913] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.791053] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.791202] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.791350] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.791551] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.791704] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.791878] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.792054] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.792235] env[62952]: DEBUG nova.virt.hardware [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.793084] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316d3eea-e963-4459-b7e5-02c9419c04cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.801361] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537a7c4d-f606-4159-aa8a-8c3e9a6b578a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.822010] env[62952]: ERROR nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Traceback (most recent call last): [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] yield resources [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self.driver.spawn(context, instance, image_meta, [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] vm_ref = self.build_virtual_machine(instance, [ 694.822010] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] for vif in network_info: [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] return self._sync_wrapper(fn, *args, **kwargs) [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self.wait() [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self[:] = self._gt.wait() [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] return self._exit_event.wait() [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.822380] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] current.throw(*self._exc) [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] result = function(*args, **kwargs) [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] return func(*args, **kwargs) [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] raise e [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] nwinfo = self.network_api.allocate_for_instance( [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] created_port_ids = self._update_ports_for_instance( [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] with excutils.save_and_reraise_exception(): [ 694.822745] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self.force_reraise() [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] raise self.value [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] updated_port = self._update_port( [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] _ensure_no_port_binding_failure(port) [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] raise exception.PortBindingFailed(port_id=port['id']) [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 694.823140] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] [ 694.823140] env[62952]: INFO nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Terminating instance [ 694.828207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquiring lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.828207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquired lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.828207] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.193987] env[62952]: DEBUG nova.network.neutron [-] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.254035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.254597] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.259020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.328s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.349302] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.428497] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.696820] env[62952]: INFO nova.compute.manager [-] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Took 1.02 seconds to deallocate network for instance. [ 695.699471] env[62952]: DEBUG nova.compute.claims [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.699685] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.766922] env[62952]: DEBUG nova.compute.utils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.769014] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.769246] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 695.828490] env[62952]: DEBUG nova.policy [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f24a30e1160a43d5be6cde39b8f79edd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '094ca56818604bee9c1f75d1493ab3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.932675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Releasing lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.933225] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.933536] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 695.934335] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23c76ee2-5776-4540-8182-f261b2b97f8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.943652] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d736865-29ab-4cdf-b727-bb22cd82db60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.969455] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64044aea-4572-48b1-93b3-93f079a829dd could not be found. [ 695.969828] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 695.970154] env[62952]: INFO nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 695.970520] env[62952]: DEBUG oslo.service.loopingcall [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.973249] env[62952]: DEBUG nova.compute.manager [-] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.973434] env[62952]: DEBUG nova.network.neutron [-] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.993619] env[62952]: DEBUG nova.network.neutron [-] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.173414] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d7cfdc-3985-4d5b-9d66-c148ccfaa48c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.181949] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d271765-3fb0-4774-a30d-4fa52b350894 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.216766] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfc9900-f7b1-466d-a6bf-768408ebdfa2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.224705] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e9b50e-26f3-4c1c-88e5-e9c17c4c9482 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.238673] env[62952]: DEBUG nova.compute.provider_tree [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.273118] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.323415] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Successfully created port: 72c35de6-bf96-4496-8bb9-8cc5f8c45497 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.348984] env[62952]: DEBUG nova.compute.manager [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Received event network-changed-003856df-04d1-443c-968c-8e4f6c633926 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.349656] env[62952]: DEBUG nova.compute.manager [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Refreshing instance network info cache due to event network-changed-003856df-04d1-443c-968c-8e4f6c633926. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 696.349656] env[62952]: DEBUG oslo_concurrency.lockutils [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] Acquiring lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.349656] env[62952]: DEBUG oslo_concurrency.lockutils [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] Acquired lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.350023] env[62952]: DEBUG nova.network.neutron [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Refreshing network info cache for port 003856df-04d1-443c-968c-8e4f6c633926 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.494600] env[62952]: DEBUG nova.network.neutron [-] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.743093] env[62952]: DEBUG nova.scheduler.client.report [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.881045] env[62952]: DEBUG nova.network.neutron [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.999822] env[62952]: INFO nova.compute.manager [-] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Took 1.03 seconds to deallocate network for instance. [ 697.003745] env[62952]: DEBUG nova.compute.claims [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.004026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.103241] env[62952]: DEBUG nova.network.neutron [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.247896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.990s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.248559] env[62952]: ERROR nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] Traceback (most recent call last): [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self.driver.spawn(context, instance, image_meta, [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] vm_ref = self.build_virtual_machine(instance, [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.248559] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] for vif in network_info: [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return self._sync_wrapper(fn, *args, **kwargs) [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self.wait() [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self[:] = self._gt.wait() [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return self._exit_event.wait() [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] result = hub.switch() [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.248941] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return self.greenlet.switch() [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] result = function(*args, **kwargs) [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] return func(*args, **kwargs) [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] raise e [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] nwinfo = self.network_api.allocate_for_instance( [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] created_port_ids = self._update_ports_for_instance( [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] with excutils.save_and_reraise_exception(): [ 697.249334] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] self.force_reraise() [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] raise self.value [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] updated_port = self._update_port( [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] _ensure_no_port_binding_failure(port) [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] raise exception.PortBindingFailed(port_id=port['id']) [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] nova.exception.PortBindingFailed: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. [ 697.249707] env[62952]: ERROR nova.compute.manager [instance: 288918a3-85be-416d-99fc-fd3450029749] [ 697.250047] env[62952]: DEBUG nova.compute.utils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.252052] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Build of instance 288918a3-85be-416d-99fc-fd3450029749 was re-scheduled: Binding failed for port 06eef808-37db-449f-835e-c20e03e67bcb, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.253174] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 697.253496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.253760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquired lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.253937] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.255928] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.755s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.280600] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.309439] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.309439] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.309439] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.309639] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.309730] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.309900] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.315832] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.315832] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.315832] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.315832] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.315832] env[62952]: DEBUG nova.virt.hardware [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.316017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2fd297-2b17-4841-bed3-97cf064923c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.327359] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e987c8-459a-4e95-897d-9c45ef5c1fe4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.606657] env[62952]: DEBUG oslo_concurrency.lockutils [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] Releasing lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.606913] env[62952]: DEBUG nova.compute.manager [req-22e19212-1957-4ae2-bcbb-1566bed842f8 req-537dafd5-8a8d-43a0-bef6-f3528269aa08 service nova] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Received event network-vif-deleted-003856df-04d1-443c-968c-8e4f6c633926 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.790564] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.880692] env[62952]: ERROR nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 697.880692] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.880692] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.880692] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.880692] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.880692] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.880692] env[62952]: ERROR nova.compute.manager raise self.value [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.880692] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.880692] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.880692] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.881192] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.881192] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.881192] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 697.881192] env[62952]: ERROR nova.compute.manager [ 697.881192] env[62952]: Traceback (most recent call last): [ 697.881192] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.881192] env[62952]: listener.cb(fileno) [ 697.881192] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.881192] env[62952]: result = function(*args, **kwargs) [ 697.881192] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.881192] env[62952]: return func(*args, **kwargs) [ 697.881192] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.881192] env[62952]: raise e [ 697.881192] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.881192] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 697.881192] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.881192] env[62952]: created_port_ids = self._update_ports_for_instance( [ 697.881192] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.881192] env[62952]: with excutils.save_and_reraise_exception(): [ 697.881192] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.881192] env[62952]: self.force_reraise() [ 697.881192] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.881192] env[62952]: raise self.value [ 697.881192] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.881192] env[62952]: updated_port = self._update_port( [ 697.881192] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.881192] env[62952]: _ensure_no_port_binding_failure(port) [ 697.881192] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.881192] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.882105] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 697.882105] env[62952]: Removing descriptor: 19 [ 697.882105] env[62952]: ERROR nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Traceback (most recent call last): [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] yield resources [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self.driver.spawn(context, instance, image_meta, [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.882105] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] vm_ref = self.build_virtual_machine(instance, [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] for vif in network_info: [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return self._sync_wrapper(fn, *args, **kwargs) [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self.wait() [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self[:] = self._gt.wait() [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return self._exit_event.wait() [ 697.882466] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] result = hub.switch() [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return self.greenlet.switch() [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] result = function(*args, **kwargs) [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return func(*args, **kwargs) [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] raise e [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] nwinfo = self.network_api.allocate_for_instance( [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.882881] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] created_port_ids = self._update_ports_for_instance( [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] with excutils.save_and_reraise_exception(): [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self.force_reraise() [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] raise self.value [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] updated_port = self._update_port( [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] _ensure_no_port_binding_failure(port) [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.883209] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] raise exception.PortBindingFailed(port_id=port['id']) [ 697.883527] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 697.883527] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] [ 697.883527] env[62952]: INFO nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Terminating instance [ 697.885187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.885354] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.885524] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.965879] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.244263] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecd821b-3f11-429a-9373-ad4555820be6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.251869] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffec7407-e5d3-4868-befb-7a1ef96bcd27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.281837] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204bf9ed-6d15-4941-8c73-3feeba83055a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.289224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f321d838-2082-4584-a80c-9c6ec4462d2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.302360] env[62952]: DEBUG nova.compute.provider_tree [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.395676] env[62952]: DEBUG nova.compute.manager [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Received event network-changed-72c35de6-bf96-4496-8bb9-8cc5f8c45497 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.395917] env[62952]: DEBUG nova.compute.manager [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Refreshing instance network info cache due to event network-changed-72c35de6-bf96-4496-8bb9-8cc5f8c45497. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 698.396130] env[62952]: DEBUG oslo_concurrency.lockutils [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] Acquiring lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.409939] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.468049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Releasing lock "refresh_cache-288918a3-85be-416d-99fc-fd3450029749" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.468889] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.468889] env[62952]: DEBUG nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.468889] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.491548] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.573043] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.805823] env[62952]: DEBUG nova.scheduler.client.report [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.995586] env[62952]: DEBUG nova.network.neutron [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.076030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.076449] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.076872] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.077743] env[62952]: DEBUG oslo_concurrency.lockutils [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] Acquired lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.077743] env[62952]: DEBUG nova.network.neutron [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Refreshing network info cache for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 699.078189] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93b1f89a-242a-4df9-ae1e-fa5189357192 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.087590] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc9d5ee-5470-48b7-addc-716330372c9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.112039] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance afd7d381-6eb1-4e5a-aaa5-34ce77953b1b could not be found. [ 699.112149] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.112837] env[62952]: INFO nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 699.112837] env[62952]: DEBUG oslo.service.loopingcall [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.112837] env[62952]: DEBUG nova.compute.manager [-] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.112837] env[62952]: DEBUG nova.network.neutron [-] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 699.132999] env[62952]: DEBUG nova.network.neutron [-] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.312385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.312385] env[62952]: ERROR nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] Traceback (most recent call last): [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self.driver.spawn(context, instance, image_meta, [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.312385] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] vm_ref = self.build_virtual_machine(instance, [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] for vif in network_info: [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return self._sync_wrapper(fn, *args, **kwargs) [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self.wait() [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self[:] = self._gt.wait() [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return self._exit_event.wait() [ 699.312801] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] result = hub.switch() [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return self.greenlet.switch() [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] result = function(*args, **kwargs) [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] return func(*args, **kwargs) [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] raise e [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] nwinfo = self.network_api.allocate_for_instance( [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.313116] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] created_port_ids = self._update_ports_for_instance( [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] with excutils.save_and_reraise_exception(): [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] self.force_reraise() [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] raise self.value [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] updated_port = self._update_port( [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] _ensure_no_port_binding_failure(port) [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.313442] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] raise exception.PortBindingFailed(port_id=port['id']) [ 699.314136] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] nova.exception.PortBindingFailed: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. [ 699.314136] env[62952]: ERROR nova.compute.manager [instance: d8653006-b429-44b5-b978-ee00a279e50e] [ 699.314136] env[62952]: DEBUG nova.compute.utils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.314136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.666s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.315392] env[62952]: INFO nova.compute.claims [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.317994] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Build of instance d8653006-b429-44b5-b978-ee00a279e50e was re-scheduled: Binding failed for port 49503947-c159-440f-ac60-05712e35899b, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.318458] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.318682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquiring lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.318829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Acquired lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.319033] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.498588] env[62952]: INFO nova.compute.manager [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 288918a3-85be-416d-99fc-fd3450029749] Took 1.03 seconds to deallocate network for instance. [ 699.605491] env[62952]: DEBUG nova.network.neutron [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.636516] env[62952]: DEBUG nova.network.neutron [-] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.734744] env[62952]: DEBUG nova.network.neutron [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.016155] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.110724] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.145137] env[62952]: INFO nova.compute.manager [-] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Took 1.03 seconds to deallocate network for instance. [ 700.147268] env[62952]: DEBUG nova.compute.claims [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 700.147459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.239343] env[62952]: DEBUG oslo_concurrency.lockutils [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] Releasing lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.239343] env[62952]: DEBUG nova.compute.manager [req-654ae7c6-e7a3-4c6e-b832-dbdf24f354c1 req-5a631302-cde7-40f0-965a-590d6398056e service nova] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Received event network-vif-deleted-72c35de6-bf96-4496-8bb9-8cc5f8c45497 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.528713] env[62952]: INFO nova.scheduler.client.report [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Deleted allocations for instance 288918a3-85be-416d-99fc-fd3450029749 [ 700.613198] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Releasing lock "refresh_cache-d8653006-b429-44b5-b978-ee00a279e50e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.613198] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.613795] env[62952]: DEBUG nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.613795] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.633551] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.755657] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44dabc84-b25c-4e89-a68a-cff25a2bede3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.768573] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fdfd1c-8a70-4216-ab3d-1f77a9ce9bc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.804565] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759923ad-a2ca-40c2-afcd-78851b5f3c3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.812356] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a004cabd-4388-481c-b3a3-d489721c8661 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.826548] env[62952]: DEBUG nova.compute.provider_tree [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.039112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96245587-26ec-4420-8603-d844bf6364ef tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "288918a3-85be-416d-99fc-fd3450029749" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.205s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.135971] env[62952]: DEBUG nova.network.neutron [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.257850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquiring lock "6f085d0c-c440-4bba-9bf8-cf2f70742748" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.258096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Lock "6f085d0c-c440-4bba-9bf8-cf2f70742748" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.330096] env[62952]: DEBUG nova.scheduler.client.report [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.542688] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.638788] env[62952]: INFO nova.compute.manager [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] [instance: d8653006-b429-44b5-b978-ee00a279e50e] Took 1.03 seconds to deallocate network for instance. [ 701.836021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.836021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.980s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.067382] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.341523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquiring lock "3c968be0-a3b6-4c00-bdfb-47df61e02b5a" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.341739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "3c968be0-a3b6-4c00-bdfb-47df61e02b5a" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.542968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "95c07667-e9d8-4cc3-b135-f80d196d179f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.543248] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "95c07667-e9d8-4cc3-b135-f80d196d179f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.665410] env[62952]: INFO nova.scheduler.client.report [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Deleted allocations for instance d8653006-b429-44b5-b978-ee00a279e50e [ 702.848530] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "3c968be0-a3b6-4c00-bdfb-47df61e02b5a" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.849188] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 702.870636] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 8db47d06-230f-4c5e-8bd6-9d5b543455b6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 702.870801] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance f54e2a1f-987b-40ac-8197-d6e4ad3ba18f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 702.870931] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 40b6590d-44e1-4d02-b1fa-46c1e6c861ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 702.871065] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance cb35ba14-6a78-4381-9c93-f2fa285468d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 702.871183] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 64044aea-4572-48b1-93b3-93f079a829dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 702.871298] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance afd7d381-6eb1-4e5a-aaa5-34ce77953b1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 702.871412] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 710a4cb2-a1f6-461e-ac26-d1e86cdd435d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 703.173064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0331be1c-076b-4aa4-999f-243dfa55c9bf tempest-DeleteServersAdminTestJSON-1224407996 tempest-DeleteServersAdminTestJSON-1224407996-project-member] Lock "d8653006-b429-44b5-b978-ee00a279e50e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.309s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.356676] env[62952]: DEBUG nova.compute.utils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.358489] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.358664] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 703.374669] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance de2daf8c-73c1-44fe-bb01-2b54f1f1c419 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.413024] env[62952]: DEBUG nova.policy [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8762d7e2e0404fd2b1c15c6f438bbe5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '608687386ecf4cedac065d9f2c5c430d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 703.676029] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.759700] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Successfully created port: 321212a8-32fe-47ed-b0e8-c5d1f8eecf06 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 703.861699] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.877089] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 244c71ad-130d-44ad-9998-3f5c51518482 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.202098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.383888] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a224dd44-bda2-464f-b946-edbe167927b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.575689] env[62952]: DEBUG nova.compute.manager [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Received event network-changed-321212a8-32fe-47ed-b0e8-c5d1f8eecf06 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.575893] env[62952]: DEBUG nova.compute.manager [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Refreshing instance network info cache due to event network-changed-321212a8-32fe-47ed-b0e8-c5d1f8eecf06. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 704.576121] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] Acquiring lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.576266] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] Acquired lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.576436] env[62952]: DEBUG nova.network.neutron [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Refreshing network info cache for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.755950] env[62952]: ERROR nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 704.755950] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.755950] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.755950] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.755950] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.755950] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.755950] env[62952]: ERROR nova.compute.manager raise self.value [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.755950] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.755950] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.755950] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.756734] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.756734] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.756734] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 704.756734] env[62952]: ERROR nova.compute.manager [ 704.756734] env[62952]: Traceback (most recent call last): [ 704.756734] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.756734] env[62952]: listener.cb(fileno) [ 704.756734] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.756734] env[62952]: result = function(*args, **kwargs) [ 704.756734] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.756734] env[62952]: return func(*args, **kwargs) [ 704.756734] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.756734] env[62952]: raise e [ 704.756734] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.756734] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 704.756734] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.756734] env[62952]: created_port_ids = self._update_ports_for_instance( [ 704.756734] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.756734] env[62952]: with excutils.save_and_reraise_exception(): [ 704.756734] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.756734] env[62952]: self.force_reraise() [ 704.756734] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.756734] env[62952]: raise self.value [ 704.756734] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.756734] env[62952]: updated_port = self._update_port( [ 704.756734] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.756734] env[62952]: _ensure_no_port_binding_failure(port) [ 704.756734] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.756734] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.757516] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 704.757516] env[62952]: Removing descriptor: 15 [ 704.871693] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.886761] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 869c5e7f-b937-4add-ace2-7d9b178c2acc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.897505] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.897648] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.897918] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.898277] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.898341] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.898517] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.898684] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.898909] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.899019] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.899197] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.899370] env[62952]: DEBUG nova.virt.hardware [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.900316] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74576875-49ca-4565-8c3c-97b85d9cd894 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.909116] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fabe2c32-066a-4e34-8dc5-6c453580642b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.925197] env[62952]: ERROR nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Traceback (most recent call last): [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] yield resources [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self.driver.spawn(context, instance, image_meta, [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] vm_ref = self.build_virtual_machine(instance, [ 704.925197] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] for vif in network_info: [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] return self._sync_wrapper(fn, *args, **kwargs) [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self.wait() [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self[:] = self._gt.wait() [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] return self._exit_event.wait() [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 704.925590] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] current.throw(*self._exc) [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] result = function(*args, **kwargs) [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] return func(*args, **kwargs) [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] raise e [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] nwinfo = self.network_api.allocate_for_instance( [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] created_port_ids = self._update_ports_for_instance( [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] with excutils.save_and_reraise_exception(): [ 704.925895] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self.force_reraise() [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] raise self.value [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] updated_port = self._update_port( [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] _ensure_no_port_binding_failure(port) [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] raise exception.PortBindingFailed(port_id=port['id']) [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 704.926207] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] [ 704.926207] env[62952]: INFO nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Terminating instance [ 704.927317] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquiring lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.096945] env[62952]: DEBUG nova.network.neutron [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.185025] env[62952]: DEBUG nova.network.neutron [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.390892] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ac386ddb-c812-4689-b9b4-b98d26ad1831 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.687207] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a7a878c-19b2-49d4-ae18-fb2d98b61998 req-7836a652-1a34-4452-91be-83f678a84e65 service nova] Releasing lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.687669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquired lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.687871] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.893355] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance f4ec1e94-1778-4296-989d-782700c6faf8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 706.209518] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.306981] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.396868] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 706.622192] env[62952]: DEBUG nova.compute.manager [req-2eacb434-fbbf-4fe7-8de1-2ab218d5d188 req-6985b63e-d77a-4dc0-9062-3c350abae143 service nova] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Received event network-vif-deleted-321212a8-32fe-47ed-b0e8-c5d1f8eecf06 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.810645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Releasing lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.811205] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.811401] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.811705] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce23c168-243f-4459-80c5-bdfb3beef5da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.821086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd951a9b-36c6-4d68-b1ca-1bcbf0586754 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.842292] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 710a4cb2-a1f6-461e-ac26-d1e86cdd435d could not be found. [ 706.842520] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.842704] env[62952]: INFO nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 706.842974] env[62952]: DEBUG oslo.service.loopingcall [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.843505] env[62952]: DEBUG nova.compute.manager [-] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.843505] env[62952]: DEBUG nova.network.neutron [-] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.862973] env[62952]: DEBUG nova.network.neutron [-] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.900247] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a825d324-0b26-4051-bdfa-4a4246f6d4e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 707.365886] env[62952]: DEBUG nova.network.neutron [-] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.403130] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 707.869805] env[62952]: INFO nova.compute.manager [-] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Took 1.03 seconds to deallocate network for instance. [ 707.872273] env[62952]: DEBUG nova.compute.claims [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.872482] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.906904] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 708.410676] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 38e05fe2-d938-43d6-ab2b-b99fb7be51d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 708.914071] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 5283bf79-bf20-4a38-912f-ced401fbd691 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 709.417805] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance fc5b891b-8432-4b2f-97d9-c93ca0a18783 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 709.924123] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 9567865d-41ae-4379-8313-e87bfc92fba8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 710.429027] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 0f8162d7-c5e6-4c66-8a10-bb06998738b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 710.934388] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 65c1a63f-4645-40fe-b429-00923e60d1cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 711.438613] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 2222309b-ce4e-46eb-a620-34553f644947 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 711.943225] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 74361d48-0a82-453a-a4a1-45173c49e901 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.447663] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 91660994-feca-40d8-8b75-d3c79110a19e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.950475] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance c7771b47-cbe8-4696-9a21-9204501c8877 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 713.453063] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4be2f0a4-5f57-4a80-9450-6d61d344799e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 713.956680] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 6f085d0c-c440-4bba-9bf8-cf2f70742748 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 714.460358] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 95c07667-e9d8-4cc3-b135-f80d196d179f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 714.460654] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 714.460792] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 714.824150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d5d112-eb96-4c74-991c-458a04a74d27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.831859] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597d8179-ee10-4193-a613-a8fd2fa4e903 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.861504] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41218d58-fca2-49a8-bebe-6813ea5e675d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.868452] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2189edac-65b6-4e00-9032-f20dff087e5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.881180] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.384093] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.888610] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 715.888951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.053s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.889146] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.885s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.717767] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779ce5b1-74fd-4887-a434-f1bc95785983 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.725285] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f728bfc8-d941-4286-8240-90e691916119 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.758867] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bae5c95-3cd5-4574-8d5a-7b4020b09651 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.766401] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fae868-2fd5-424a-93ba-d080a213b092 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.779385] env[62952]: DEBUG nova.compute.provider_tree [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.284392] env[62952]: DEBUG nova.scheduler.client.report [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.788540] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.899s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.789197] env[62952]: ERROR nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Traceback (most recent call last): [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self.driver.spawn(context, instance, image_meta, [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] vm_ref = self.build_virtual_machine(instance, [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.789197] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] for vif in network_info: [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] return self._sync_wrapper(fn, *args, **kwargs) [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self.wait() [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self[:] = self._gt.wait() [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] return self._exit_event.wait() [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] current.throw(*self._exc) [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.789490] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] result = function(*args, **kwargs) [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] return func(*args, **kwargs) [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] raise e [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] nwinfo = self.network_api.allocate_for_instance( [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] created_port_ids = self._update_ports_for_instance( [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] with excutils.save_and_reraise_exception(): [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] self.force_reraise() [ 717.789781] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] raise self.value [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] updated_port = self._update_port( [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] _ensure_no_port_binding_failure(port) [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] raise exception.PortBindingFailed(port_id=port['id']) [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] nova.exception.PortBindingFailed: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. [ 717.790095] env[62952]: ERROR nova.compute.manager [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] [ 717.790095] env[62952]: DEBUG nova.compute.utils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.791301] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.367s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.794310] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Build of instance 8db47d06-230f-4c5e-8bd6-9d5b543455b6 was re-scheduled: Binding failed for port c2cb65ec-f4c4-43ee-abfe-11c6c590e929, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.794522] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.794752] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquiring lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.794900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Acquired lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.795071] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.317541] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.401539] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.734293] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df884e4-e03d-457b-9b16-bb2a6fe7eaab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.742223] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02af21ba-2a1e-4cc8-9116-f17dc11c999a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.772118] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d374ed2f-20cb-4bad-b78b-eb5ed79bad7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.779145] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d751072-5cb6-4fcb-adbf-8cd09788c359 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.791886] env[62952]: DEBUG nova.compute.provider_tree [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.903379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Releasing lock "refresh_cache-8db47d06-230f-4c5e-8bd6-9d5b543455b6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.903633] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.903822] env[62952]: DEBUG nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.903991] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.919134] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.295899] env[62952]: DEBUG nova.scheduler.client.report [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.421076] env[62952]: DEBUG nova.network.neutron [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.801200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.801767] env[62952]: ERROR nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Traceback (most recent call last): [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self.driver.spawn(context, instance, image_meta, [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] vm_ref = self.build_virtual_machine(instance, [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.801767] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] for vif in network_info: [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return self._sync_wrapper(fn, *args, **kwargs) [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self.wait() [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self[:] = self._gt.wait() [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return self._exit_event.wait() [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] result = hub.switch() [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 719.802261] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return self.greenlet.switch() [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] result = function(*args, **kwargs) [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] return func(*args, **kwargs) [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] raise e [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] nwinfo = self.network_api.allocate_for_instance( [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] created_port_ids = self._update_ports_for_instance( [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] with excutils.save_and_reraise_exception(): [ 719.802561] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] self.force_reraise() [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] raise self.value [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] updated_port = self._update_port( [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] _ensure_no_port_binding_failure(port) [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] raise exception.PortBindingFailed(port_id=port['id']) [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] nova.exception.PortBindingFailed: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. [ 719.802869] env[62952]: ERROR nova.compute.manager [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] [ 719.803154] env[62952]: DEBUG nova.compute.utils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.803701] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.752s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.806872] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Build of instance f54e2a1f-987b-40ac-8197-d6e4ad3ba18f was re-scheduled: Binding failed for port f09338d9-1abf-4bf6-bc1a-d40ad7ff3d0b, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.807305] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.807524] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquiring lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.807670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Acquired lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.807867] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.924465] env[62952]: INFO nova.compute.manager [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] [instance: 8db47d06-230f-4c5e-8bd6-9d5b543455b6] Took 1.02 seconds to deallocate network for instance. [ 720.330961] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.376328] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.643107] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2bfd2b-02d4-403b-878d-60cee0c1b207 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.650365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01011964-9346-4266-aad6-6bd9a34a7650 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.679950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e113c55d-adf3-47b0-968c-9bfed76a88f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.686673] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e12a9d-46b7-4b0b-9b87-b4c46d4394fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.699374] env[62952]: DEBUG nova.compute.provider_tree [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.878830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Releasing lock "refresh_cache-f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.879120] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.879295] env[62952]: DEBUG nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.879464] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.894639] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.951428] env[62952]: INFO nova.scheduler.client.report [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Deleted allocations for instance 8db47d06-230f-4c5e-8bd6-9d5b543455b6 [ 721.202518] env[62952]: DEBUG nova.scheduler.client.report [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.399211] env[62952]: DEBUG nova.network.neutron [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.459786] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4de5d918-a790-4830-b328-be50e548befd tempest-InstanceActionsV221TestJSON-1361617694 tempest-InstanceActionsV221TestJSON-1361617694-project-member] Lock "8db47d06-230f-4c5e-8bd6-9d5b543455b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.835s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.708050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.708394] env[62952]: ERROR nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Traceback (most recent call last): [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self.driver.spawn(context, instance, image_meta, [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] vm_ref = self.build_virtual_machine(instance, [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.708394] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] for vif in network_info: [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] return self._sync_wrapper(fn, *args, **kwargs) [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self.wait() [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self[:] = self._gt.wait() [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] return self._exit_event.wait() [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] current.throw(*self._exc) [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.708747] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] result = function(*args, **kwargs) [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] return func(*args, **kwargs) [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] raise e [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] nwinfo = self.network_api.allocate_for_instance( [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] created_port_ids = self._update_ports_for_instance( [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] with excutils.save_and_reraise_exception(): [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] self.force_reraise() [ 721.709126] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] raise self.value [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] updated_port = self._update_port( [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] _ensure_no_port_binding_failure(port) [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] raise exception.PortBindingFailed(port_id=port['id']) [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] nova.exception.PortBindingFailed: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. [ 721.709531] env[62952]: ERROR nova.compute.manager [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] [ 721.709531] env[62952]: DEBUG nova.compute.utils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.710669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.169s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.712315] env[62952]: INFO nova.compute.claims [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.715035] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Build of instance 40b6590d-44e1-4d02-b1fa-46c1e6c861ea was re-scheduled: Binding failed for port d9c5eae3-096a-4b17-a34b-0874b5bd8c36, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.715475] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.715720] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquiring lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.715875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Acquired lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.716014] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.902369] env[62952]: INFO nova.compute.manager [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] [instance: f54e2a1f-987b-40ac-8197-d6e4ad3ba18f] Took 1.02 seconds to deallocate network for instance. [ 721.966736] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.255163] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.336322] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.488583] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.839163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Releasing lock "refresh_cache-40b6590d-44e1-4d02-b1fa-46c1e6c861ea" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.839447] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.839565] env[62952]: DEBUG nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.839728] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.859925] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.937016] env[62952]: INFO nova.scheduler.client.report [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Deleted allocations for instance f54e2a1f-987b-40ac-8197-d6e4ad3ba18f [ 723.241218] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2bee4fe-ab62-4ec9-9b0a-7cc4e499cd66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.249165] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b10fec-fff4-404a-852a-fe53e5a92562 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.281263] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4d3f24-2b21-44c5-b279-1e655b746b25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.290034] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52acf906-4d06-4957-8cd3-80dd8001473e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.304541] env[62952]: DEBUG nova.compute.provider_tree [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.365814] env[62952]: DEBUG nova.network.neutron [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.445323] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f33e9603-1122-4787-a4f0-cf7618027a6a tempest-ServerDiagnosticsTest-1080211335 tempest-ServerDiagnosticsTest-1080211335-project-member] Lock "f54e2a1f-987b-40ac-8197-d6e4ad3ba18f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.258s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.825638] env[62952]: ERROR nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [req-1bb45ec5-dac1-42e1-ba55-a6815559a185] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1bb45ec5-dac1-42e1-ba55-a6815559a185"}]} [ 723.841750] env[62952]: DEBUG nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 723.855043] env[62952]: DEBUG nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 723.855330] env[62952]: DEBUG nova.compute.provider_tree [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.866790] env[62952]: DEBUG nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 723.868809] env[62952]: INFO nova.compute.manager [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] [instance: 40b6590d-44e1-4d02-b1fa-46c1e6c861ea] Took 1.03 seconds to deallocate network for instance. [ 723.888190] env[62952]: DEBUG nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 723.948666] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.237694] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a9cc17-33a3-487c-989c-b222e05eb634 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.245692] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6202298-d998-4f4a-86d6-46e1145e4905 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.279902] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf01da15-6566-48fd-806c-ab2c6424d027 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.288114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bf92fc-fd17-4f63-a493-8186d36a90d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.303839] env[62952]: DEBUG nova.compute.provider_tree [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.466722] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.842054] env[62952]: DEBUG nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 60 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 724.842333] env[62952]: DEBUG nova.compute.provider_tree [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 60 to 61 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 724.842512] env[62952]: DEBUG nova.compute.provider_tree [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.904113] env[62952]: INFO nova.scheduler.client.report [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Deleted allocations for instance 40b6590d-44e1-4d02-b1fa-46c1e6c861ea [ 725.350717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.640s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.351307] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 725.353888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.654s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.409270] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0c9a9821-e78c-4b0e-9901-8a8b0512a171 tempest-MigrationsAdminTest-674001853 tempest-MigrationsAdminTest-674001853-project-member] Lock "40b6590d-44e1-4d02-b1fa-46c1e6c861ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.184s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.843862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquiring lock "ca7fc13d-e9be-481c-ae58-78377627800c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.844150] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Lock "ca7fc13d-e9be-481c-ae58-78377627800c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.857962] env[62952]: DEBUG nova.compute.utils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.865025] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.865025] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 725.912744] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.933181] env[62952]: DEBUG nova.policy [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fda30da3f0e7444b9a109fd15c1bbb20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7a9b722dee442b1bd6fa656a3689fa0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.241465] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e17129-4669-463f-a27c-79d2dfd6f0e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.249166] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b959b83-15fb-4d9c-a1dc-3bce71a8a9c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.279575] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Successfully created port: d208e472-8506-4ab4-8e8a-a63b5e63a3d9 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.281895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e45e80-a13d-4d2c-b5b9-c4708913dfb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.289104] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b40b63-f785-4df6-8aa8-a687d991b018 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.315771] env[62952]: DEBUG nova.compute.provider_tree [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 726.363488] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.436144] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.834368] env[62952]: ERROR nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [req-ff6920e3-6279-46fb-82a5-47b7f7de2d9c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ff6920e3-6279-46fb-82a5-47b7f7de2d9c"}]}: nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 726.860895] env[62952]: DEBUG nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 726.877525] env[62952]: DEBUG nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 726.877748] env[62952]: DEBUG nova.compute.provider_tree [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 726.889909] env[62952]: DEBUG nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 726.911321] env[62952]: DEBUG nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 727.107154] env[62952]: DEBUG nova.compute.manager [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Received event network-changed-d208e472-8506-4ab4-8e8a-a63b5e63a3d9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.107154] env[62952]: DEBUG nova.compute.manager [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Refreshing instance network info cache due to event network-changed-d208e472-8506-4ab4-8e8a-a63b5e63a3d9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.107154] env[62952]: DEBUG oslo_concurrency.lockutils [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] Acquiring lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.107154] env[62952]: DEBUG oslo_concurrency.lockutils [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] Acquired lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.107154] env[62952]: DEBUG nova.network.neutron [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Refreshing network info cache for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.290946] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21042d6-3f48-4e6e-8349-62d2101db383 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.298939] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048d071c-1014-4119-a1d4-89a3ac5d1f6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.329659] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5625c0d0-1162-43c4-99f6-ae422c18b06c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.337214] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6977a27-7e96-48a6-930e-4775e6caf512 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.350227] env[62952]: DEBUG nova.compute.provider_tree [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 727.374184] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.401322] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.401580] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.401735] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.401937] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.402064] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.402220] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.402462] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.402605] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.402768] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.402948] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.403131] env[62952]: DEBUG nova.virt.hardware [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.403961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ca2705-d39a-4239-944b-cd43b873ba6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.411342] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371594b6-ef13-4752-9816-495ba487d02d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.482771] env[62952]: ERROR nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 727.482771] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.482771] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.482771] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.482771] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.482771] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.482771] env[62952]: ERROR nova.compute.manager raise self.value [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.482771] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.482771] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.482771] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.483354] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.483354] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.483354] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 727.483354] env[62952]: ERROR nova.compute.manager [ 727.483354] env[62952]: Traceback (most recent call last): [ 727.483354] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.483354] env[62952]: listener.cb(fileno) [ 727.483354] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.483354] env[62952]: result = function(*args, **kwargs) [ 727.483354] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.483354] env[62952]: return func(*args, **kwargs) [ 727.483354] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.483354] env[62952]: raise e [ 727.483354] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.483354] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 727.483354] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.483354] env[62952]: created_port_ids = self._update_ports_for_instance( [ 727.483354] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.483354] env[62952]: with excutils.save_and_reraise_exception(): [ 727.483354] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.483354] env[62952]: self.force_reraise() [ 727.483354] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.483354] env[62952]: raise self.value [ 727.483354] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.483354] env[62952]: updated_port = self._update_port( [ 727.483354] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.483354] env[62952]: _ensure_no_port_binding_failure(port) [ 727.483354] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.483354] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.484271] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 727.484271] env[62952]: Removing descriptor: 15 [ 727.484271] env[62952]: ERROR nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Traceback (most recent call last): [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] yield resources [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self.driver.spawn(context, instance, image_meta, [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.484271] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] vm_ref = self.build_virtual_machine(instance, [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] for vif in network_info: [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return self._sync_wrapper(fn, *args, **kwargs) [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self.wait() [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self[:] = self._gt.wait() [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return self._exit_event.wait() [ 727.484827] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] result = hub.switch() [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return self.greenlet.switch() [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] result = function(*args, **kwargs) [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return func(*args, **kwargs) [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] raise e [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] nwinfo = self.network_api.allocate_for_instance( [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.485367] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] created_port_ids = self._update_ports_for_instance( [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] with excutils.save_and_reraise_exception(): [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self.force_reraise() [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] raise self.value [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] updated_port = self._update_port( [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] _ensure_no_port_binding_failure(port) [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.485773] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] raise exception.PortBindingFailed(port_id=port['id']) [ 727.486294] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 727.486294] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] [ 727.486294] env[62952]: INFO nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Terminating instance [ 727.486294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.635311] env[62952]: DEBUG nova.network.neutron [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.756153] env[62952]: DEBUG nova.network.neutron [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.895267] env[62952]: DEBUG nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 62 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 727.895546] env[62952]: DEBUG nova.compute.provider_tree [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 62 to 63 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 727.895707] env[62952]: DEBUG nova.compute.provider_tree [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 728.259566] env[62952]: DEBUG oslo_concurrency.lockutils [req-b75ef6c0-bced-4d8c-a8f4-3add28cd952b req-79721b9e-b26e-4ed6-8cf3-4aabb248f0c1 service nova] Releasing lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.259999] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquired lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.260204] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.401426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.047s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.402371] env[62952]: ERROR nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Traceback (most recent call last): [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self.driver.spawn(context, instance, image_meta, [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] vm_ref = self.build_virtual_machine(instance, [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.402371] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] for vif in network_info: [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return self._sync_wrapper(fn, *args, **kwargs) [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self.wait() [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self[:] = self._gt.wait() [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return self._exit_event.wait() [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] result = hub.switch() [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.402770] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return self.greenlet.switch() [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] result = function(*args, **kwargs) [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] return func(*args, **kwargs) [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] raise e [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] nwinfo = self.network_api.allocate_for_instance( [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] created_port_ids = self._update_ports_for_instance( [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] with excutils.save_and_reraise_exception(): [ 728.403134] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] self.force_reraise() [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] raise self.value [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] updated_port = self._update_port( [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] _ensure_no_port_binding_failure(port) [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] raise exception.PortBindingFailed(port_id=port['id']) [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] nova.exception.PortBindingFailed: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. [ 728.403575] env[62952]: ERROR nova.compute.manager [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] [ 728.403852] env[62952]: DEBUG nova.compute.utils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.404375] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.400s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.407571] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Build of instance cb35ba14-6a78-4381-9c93-f2fa285468d7 was re-scheduled: Binding failed for port 6789cad3-4447-4caf-a77a-4151a1974032, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.408046] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.408295] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquiring lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.408914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Acquired lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.409094] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.781318] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.899474] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.933285] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.029096] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.135932] env[62952]: DEBUG nova.compute.manager [req-0b929aab-8064-40e5-9e10-685c7795e2fb req-76bab0c7-4971-4511-9b7c-549ed9848ca7 service nova] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Received event network-vif-deleted-d208e472-8506-4ab4-8e8a-a63b5e63a3d9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.288382] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a05cd9d-79d5-4e11-8f4b-d4afbb6f9266 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.296140] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6684fa6-2cb2-4ede-b57c-372151adf88a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.325120] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15820c52-dc3e-48e0-8ec2-8d5683db6ebb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.331808] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b61d918-5463-48ea-adca-e89d2712374d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.344146] env[62952]: DEBUG nova.compute.provider_tree [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.402438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Releasing lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.402861] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.403065] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.403341] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01f78b72-a7a8-4d63-8922-7a51482e7bc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.411701] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92a4a2e-d28e-48d2-917d-01d2eeb1dd68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.432634] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de2daf8c-73c1-44fe-bb01-2b54f1f1c419 could not be found. [ 729.432833] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.433017] env[62952]: INFO nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Took 0.03 seconds to destroy the instance on the hypervisor. [ 729.433263] env[62952]: DEBUG oslo.service.loopingcall [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.433498] env[62952]: DEBUG nova.compute.manager [-] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.433582] env[62952]: DEBUG nova.network.neutron [-] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.447900] env[62952]: DEBUG nova.network.neutron [-] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.536270] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Releasing lock "refresh_cache-cb35ba14-6a78-4381-9c93-f2fa285468d7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.536270] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.536270] env[62952]: DEBUG nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.536270] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.552026] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.847334] env[62952]: DEBUG nova.scheduler.client.report [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.950674] env[62952]: DEBUG nova.network.neutron [-] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.057519] env[62952]: DEBUG nova.network.neutron [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.351851] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.947s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.352569] env[62952]: ERROR nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Traceback (most recent call last): [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self.driver.spawn(context, instance, image_meta, [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] vm_ref = self.build_virtual_machine(instance, [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.352569] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] for vif in network_info: [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] return self._sync_wrapper(fn, *args, **kwargs) [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self.wait() [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self[:] = self._gt.wait() [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] return self._exit_event.wait() [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] current.throw(*self._exc) [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.352912] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] result = function(*args, **kwargs) [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] return func(*args, **kwargs) [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] raise e [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] nwinfo = self.network_api.allocate_for_instance( [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] created_port_ids = self._update_ports_for_instance( [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] with excutils.save_and_reraise_exception(): [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] self.force_reraise() [ 730.353283] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] raise self.value [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] updated_port = self._update_port( [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] _ensure_no_port_binding_failure(port) [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] raise exception.PortBindingFailed(port_id=port['id']) [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] nova.exception.PortBindingFailed: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. [ 730.353666] env[62952]: ERROR nova.compute.manager [instance: 64044aea-4572-48b1-93b3-93f079a829dd] [ 730.353666] env[62952]: DEBUG nova.compute.utils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.355148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.208s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.358111] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Build of instance 64044aea-4572-48b1-93b3-93f079a829dd was re-scheduled: Binding failed for port 003856df-04d1-443c-968c-8e4f6c633926, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.358537] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.358754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquiring lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.358897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Acquired lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.359100] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.452885] env[62952]: INFO nova.compute.manager [-] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Took 1.02 seconds to deallocate network for instance. [ 730.455225] env[62952]: DEBUG nova.compute.claims [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 730.455403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.458911] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.459146] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.559673] env[62952]: INFO nova.compute.manager [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] [instance: cb35ba14-6a78-4381-9c93-f2fa285468d7] Took 1.02 seconds to deallocate network for instance. [ 730.877857] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.974835] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.240131] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccfc1e2-98cf-4d59-93a5-81a26fa670ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.250782] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7976750a-742d-4045-99b1-9514adf4fddd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.288839] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22096aed-0d13-41d8-a74c-e838ce6e31af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.296631] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0eb185b-4f88-4429-ab2f-3ce26d9a5dec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.310372] env[62952]: DEBUG nova.compute.provider_tree [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.478844] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Releasing lock "refresh_cache-64044aea-4572-48b1-93b3-93f079a829dd" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.479111] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.479315] env[62952]: DEBUG nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.479498] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.494851] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.587413] env[62952]: INFO nova.scheduler.client.report [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Deleted allocations for instance cb35ba14-6a78-4381-9c93-f2fa285468d7 [ 731.723577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquiring lock "4846f1d8-7628-405e-bef0-0d795a973a34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.723858] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Lock "4846f1d8-7628-405e-bef0-0d795a973a34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.815102] env[62952]: DEBUG nova.scheduler.client.report [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.996962] env[62952]: DEBUG nova.network.neutron [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.095316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1fd00242-b7de-4e3c-9925-a989db253696 tempest-ServerActionsTestJSON-1068459111 tempest-ServerActionsTestJSON-1068459111-project-member] Lock "cb35ba14-6a78-4381-9c93-f2fa285468d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.559s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.319837] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.320498] env[62952]: ERROR nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Traceback (most recent call last): [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self.driver.spawn(context, instance, image_meta, [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] vm_ref = self.build_virtual_machine(instance, [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.320498] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] for vif in network_info: [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return self._sync_wrapper(fn, *args, **kwargs) [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self.wait() [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self[:] = self._gt.wait() [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return self._exit_event.wait() [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] result = hub.switch() [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.320795] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return self.greenlet.switch() [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] result = function(*args, **kwargs) [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] return func(*args, **kwargs) [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] raise e [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] nwinfo = self.network_api.allocate_for_instance( [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] created_port_ids = self._update_ports_for_instance( [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] with excutils.save_and_reraise_exception(): [ 732.321159] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] self.force_reraise() [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] raise self.value [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] updated_port = self._update_port( [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] _ensure_no_port_binding_failure(port) [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] raise exception.PortBindingFailed(port_id=port['id']) [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] nova.exception.PortBindingFailed: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. [ 732.321512] env[62952]: ERROR nova.compute.manager [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] [ 732.321821] env[62952]: DEBUG nova.compute.utils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.322538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.255s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.324433] env[62952]: INFO nova.compute.claims [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.327014] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Build of instance afd7d381-6eb1-4e5a-aaa5-34ce77953b1b was re-scheduled: Binding failed for port 72c35de6-bf96-4496-8bb9-8cc5f8c45497, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.327467] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.327685] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.327831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.327985] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.499188] env[62952]: INFO nova.compute.manager [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] [instance: 64044aea-4572-48b1-93b3-93f079a829dd] Took 1.02 seconds to deallocate network for instance. [ 732.597659] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.860684] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.953630] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.126029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.485207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.485207] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.485207] env[62952]: DEBUG nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.485207] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.514874] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.548618] env[62952]: INFO nova.scheduler.client.report [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Deleted allocations for instance 64044aea-4572-48b1-93b3-93f079a829dd [ 733.817343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949eb682-ad91-42ec-b60c-918ca732b65f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.825046] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea01888-2630-4cb0-81a7-097c5af2bf04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.864784] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2f5882-2720-4a2a-85d4-584f4c4779e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.872427] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227433d3-515c-4ba8-9722-5dfa163a61da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.884981] env[62952]: DEBUG nova.compute.provider_tree [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.021662] env[62952]: DEBUG nova.network.neutron [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.065150] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bcf61173-4091-4fc1-ac45-d535f17e1268 tempest-ImagesOneServerNegativeTestJSON-1454207902 tempest-ImagesOneServerNegativeTestJSON-1454207902-project-member] Lock "64044aea-4572-48b1-93b3-93f079a829dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.142s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.388315] env[62952]: DEBUG nova.scheduler.client.report [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.524675] env[62952]: INFO nova.compute.manager [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: afd7d381-6eb1-4e5a-aaa5-34ce77953b1b] Took 1.07 seconds to deallocate network for instance. [ 734.569142] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.897161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.897161] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.899153] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.698s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.900653] env[62952]: INFO nova.compute.claims [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.106455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.407087] env[62952]: DEBUG nova.compute.utils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.413619] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.416237] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.490099] env[62952]: DEBUG nova.policy [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6165653e44f147ee91e4a2edef90337e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd43742a77a784ed087a0105400e61435', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.557367] env[62952]: INFO nova.scheduler.client.report [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocations for instance afd7d381-6eb1-4e5a-aaa5-34ce77953b1b [ 735.925656] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.071278] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff5f400-88e4-44bb-b9dd-edac07e8327d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "afd7d381-6eb1-4e5a-aaa5-34ce77953b1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.156s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.139804] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Successfully created port: fadeb616-d710-4ea9-897e-1f887a92b3f2 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.305577] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556dc48f-81bb-4c90-a98b-6243b0350c86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.316100] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef02c88-a9d8-4a87-91c7-125c7e8218e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.343933] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1676acd-7733-4b6b-b889-17d9e29ad804 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.351135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e119d00a-73c3-4afc-94c8-415f0d74a774 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.364503] env[62952]: DEBUG nova.compute.provider_tree [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.576605] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.867641] env[62952]: DEBUG nova.scheduler.client.report [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.934532] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.970169] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.970169] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.970169] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.970322] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.970322] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.970322] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.970322] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.970322] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.970472] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.970472] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.971517] env[62952]: DEBUG nova.virt.hardware [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.971679] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c992eebf-991b-47c9-93fc-3198b45854ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.979869] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e6f057-658b-4d9e-9892-76f49920e997 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.105680] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.169695] env[62952]: DEBUG nova.compute.manager [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Received event network-changed-fadeb616-d710-4ea9-897e-1f887a92b3f2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.169695] env[62952]: DEBUG nova.compute.manager [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Refreshing instance network info cache due to event network-changed-fadeb616-d710-4ea9-897e-1f887a92b3f2. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.169873] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] Acquiring lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.170019] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] Acquired lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.170187] env[62952]: DEBUG nova.network.neutron [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Refreshing network info cache for port fadeb616-d710-4ea9-897e-1f887a92b3f2 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 737.373419] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.373942] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.376911] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.504s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.443021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "16224e22-7091-4381-9f40-3f52f24bb724" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.443266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "16224e22-7091-4381-9f40-3f52f24bb724" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.559203] env[62952]: ERROR nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 737.559203] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.559203] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.559203] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.559203] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.559203] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.559203] env[62952]: ERROR nova.compute.manager raise self.value [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.559203] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.559203] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.559203] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.559672] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.559672] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.559672] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 737.559672] env[62952]: ERROR nova.compute.manager [ 737.559672] env[62952]: Traceback (most recent call last): [ 737.559672] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.559672] env[62952]: listener.cb(fileno) [ 737.559672] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.559672] env[62952]: result = function(*args, **kwargs) [ 737.559672] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.559672] env[62952]: return func(*args, **kwargs) [ 737.559672] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.559672] env[62952]: raise e [ 737.559672] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.559672] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 737.559672] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.559672] env[62952]: created_port_ids = self._update_ports_for_instance( [ 737.559672] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.559672] env[62952]: with excutils.save_and_reraise_exception(): [ 737.559672] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.559672] env[62952]: self.force_reraise() [ 737.559672] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.559672] env[62952]: raise self.value [ 737.559672] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.559672] env[62952]: updated_port = self._update_port( [ 737.559672] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.559672] env[62952]: _ensure_no_port_binding_failure(port) [ 737.559672] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.559672] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.560368] env[62952]: nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 737.560368] env[62952]: Removing descriptor: 15 [ 737.560368] env[62952]: ERROR nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Traceback (most recent call last): [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] yield resources [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self.driver.spawn(context, instance, image_meta, [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.560368] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] vm_ref = self.build_virtual_machine(instance, [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] for vif in network_info: [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return self._sync_wrapper(fn, *args, **kwargs) [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self.wait() [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self[:] = self._gt.wait() [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return self._exit_event.wait() [ 737.560663] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] result = hub.switch() [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return self.greenlet.switch() [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] result = function(*args, **kwargs) [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return func(*args, **kwargs) [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] raise e [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] nwinfo = self.network_api.allocate_for_instance( [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.560977] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] created_port_ids = self._update_ports_for_instance( [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] with excutils.save_and_reraise_exception(): [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self.force_reraise() [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] raise self.value [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] updated_port = self._update_port( [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] _ensure_no_port_binding_failure(port) [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.561372] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] raise exception.PortBindingFailed(port_id=port['id']) [ 737.561887] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 737.561887] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] [ 737.561887] env[62952]: INFO nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Terminating instance [ 737.562366] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.698046] env[62952]: DEBUG nova.network.neutron [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.752027] env[62952]: DEBUG nova.network.neutron [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.885385] env[62952]: DEBUG nova.compute.utils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.891190] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 737.891190] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.944255] env[62952]: DEBUG nova.policy [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '792a25833c124a5aaa78da804b045495', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4c95338402e4b73b51f8700dece8bab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.256891] env[62952]: DEBUG oslo_concurrency.lockutils [req-0c280c6b-6954-4bb1-ad23-4c08e14488be req-99ff3f70-af57-428b-9bec-e8f1617cb766 service nova] Releasing lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.257348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquired lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.257622] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.287017] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Successfully created port: aa82811b-2e65-4278-b2d4-40301d1e8b7d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.329408] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f821d94f-b0d9-41b0-aa7d-dc01e208f822 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.333857] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1c2ab5-3f01-418c-ab33-64ee7115e4db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.365114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d057ff80-9cda-47d0-a5bd-46648d03d106 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.372950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8930d42c-b6b4-40d2-9151-1bacb90ff3bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.385932] env[62952]: DEBUG nova.compute.provider_tree [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.393132] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.785280] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.872533] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.892023] env[62952]: DEBUG nova.scheduler.client.report [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.229088] env[62952]: DEBUG nova.compute.manager [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Received event network-vif-deleted-fadeb616-d710-4ea9-897e-1f887a92b3f2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.229304] env[62952]: DEBUG nova.compute.manager [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Received event network-changed-aa82811b-2e65-4278-b2d4-40301d1e8b7d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.229461] env[62952]: DEBUG nova.compute.manager [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Refreshing instance network info cache due to event network-changed-aa82811b-2e65-4278-b2d4-40301d1e8b7d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.231030] env[62952]: DEBUG oslo_concurrency.lockutils [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] Acquiring lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.231030] env[62952]: DEBUG oslo_concurrency.lockutils [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] Acquired lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.231030] env[62952]: DEBUG nova.network.neutron [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Refreshing network info cache for port aa82811b-2e65-4278-b2d4-40301d1e8b7d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.378020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Releasing lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.378020] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.378020] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.378020] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1ea8f33-d45e-4359-a204-3f0813f4e5fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.388281] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ca97c2-0740-4e4f-b44a-0b5869a259a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.403145] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.026s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.403889] env[62952]: ERROR nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Traceback (most recent call last): [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self.driver.spawn(context, instance, image_meta, [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] vm_ref = self.build_virtual_machine(instance, [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.403889] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] for vif in network_info: [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] return self._sync_wrapper(fn, *args, **kwargs) [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self.wait() [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self[:] = self._gt.wait() [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] return self._exit_event.wait() [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] current.throw(*self._exc) [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.404256] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] result = function(*args, **kwargs) [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] return func(*args, **kwargs) [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] raise e [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] nwinfo = self.network_api.allocate_for_instance( [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] created_port_ids = self._update_ports_for_instance( [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] with excutils.save_and_reraise_exception(): [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] self.force_reraise() [ 739.404664] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] raise self.value [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] updated_port = self._update_port( [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] _ensure_no_port_binding_failure(port) [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] raise exception.PortBindingFailed(port_id=port['id']) [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] nova.exception.PortBindingFailed: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. [ 739.407806] env[62952]: ERROR nova.compute.manager [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] [ 739.407806] env[62952]: DEBUG nova.compute.utils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.408172] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.409602] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Build of instance 710a4cb2-a1f6-461e-ac26-d1e86cdd435d was re-scheduled: Binding failed for port 321212a8-32fe-47ed-b0e8-c5d1f8eecf06, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.410081] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.410312] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquiring lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.410457] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Acquired lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.410621] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.412299] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.924s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.414908] env[62952]: INFO nova.compute.claims [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.431035] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 244c71ad-130d-44ad-9998-3f5c51518482 could not be found. [ 739.431035] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.431141] env[62952]: INFO nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Took 0.06 seconds to destroy the instance on the hypervisor. [ 739.431388] env[62952]: DEBUG oslo.service.loopingcall [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.434213] env[62952]: DEBUG nova.compute.manager [-] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.434213] env[62952]: DEBUG nova.network.neutron [-] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.440189] env[62952]: ERROR nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 739.440189] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.440189] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.440189] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.440189] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.440189] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.440189] env[62952]: ERROR nova.compute.manager raise self.value [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.440189] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.440189] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.440189] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.440643] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.440643] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.440643] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 739.440643] env[62952]: ERROR nova.compute.manager [ 739.440643] env[62952]: Traceback (most recent call last): [ 739.440643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.440643] env[62952]: listener.cb(fileno) [ 739.440643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.440643] env[62952]: result = function(*args, **kwargs) [ 739.440643] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.440643] env[62952]: return func(*args, **kwargs) [ 739.440643] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.440643] env[62952]: raise e [ 739.440643] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.440643] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 739.440643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.440643] env[62952]: created_port_ids = self._update_ports_for_instance( [ 739.440643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.440643] env[62952]: with excutils.save_and_reraise_exception(): [ 739.440643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.440643] env[62952]: self.force_reraise() [ 739.440643] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.440643] env[62952]: raise self.value [ 739.440643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.440643] env[62952]: updated_port = self._update_port( [ 739.440643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.440643] env[62952]: _ensure_no_port_binding_failure(port) [ 739.440643] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.440643] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.442273] env[62952]: nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 739.442273] env[62952]: Removing descriptor: 15 [ 739.445684] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.445809] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.446046] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.446129] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.446273] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.446417] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.446616] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.446769] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.446934] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.447108] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.447280] env[62952]: DEBUG nova.virt.hardware [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.448129] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca55c09d-71c2-4628-a57b-df89ad8f0895 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.451896] env[62952]: DEBUG nova.network.neutron [-] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.460714] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d88c2a-68cf-4654-86bc-6f3fe920e9f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.476567] env[62952]: ERROR nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] Traceback (most recent call last): [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] yield resources [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self.driver.spawn(context, instance, image_meta, [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] vm_ref = self.build_virtual_machine(instance, [ 739.476567] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] for vif in network_info: [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] return self._sync_wrapper(fn, *args, **kwargs) [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self.wait() [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self[:] = self._gt.wait() [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] return self._exit_event.wait() [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.476932] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] current.throw(*self._exc) [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] result = function(*args, **kwargs) [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] return func(*args, **kwargs) [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] raise e [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] nwinfo = self.network_api.allocate_for_instance( [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] created_port_ids = self._update_ports_for_instance( [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] with excutils.save_and_reraise_exception(): [ 739.477331] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self.force_reraise() [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] raise self.value [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] updated_port = self._update_port( [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] _ensure_no_port_binding_failure(port) [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] raise exception.PortBindingFailed(port_id=port['id']) [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 739.477695] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] [ 739.477695] env[62952]: INFO nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Terminating instance [ 739.479736] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquiring lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.753165] env[62952]: DEBUG nova.network.neutron [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.863564] env[62952]: DEBUG nova.network.neutron [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.947653] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.954263] env[62952]: DEBUG nova.network.neutron [-] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.045744] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.350659] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquiring lock "b4ea4cff-ada0-4355-a514-d8b179616535" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.350869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Lock "b4ea4cff-ada0-4355-a514-d8b179616535" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.366161] env[62952]: DEBUG oslo_concurrency.lockutils [req-740472d0-8603-434d-a9f5-a336d6c9c117 req-95252a82-15c7-40cb-ad0e-c304d492f68b service nova] Releasing lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.366737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquired lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.366921] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.457432] env[62952]: INFO nova.compute.manager [-] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Took 1.02 seconds to deallocate network for instance. [ 740.464574] env[62952]: DEBUG nova.compute.claims [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.464761] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.547872] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Releasing lock "refresh_cache-710a4cb2-a1f6-461e-ac26-d1e86cdd435d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.548125] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.548280] env[62952]: DEBUG nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.548439] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.562798] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.852228] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574a9d20-ceb2-4de8-b652-2abaa64ec95a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.864152] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bc6c89-f97e-49fb-92f4-49ce368ea987 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.905201] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9f9f38-704d-4b98-930f-2b8a01ed22b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.913114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20fe00a-8749-4716-8ab5-8008d0ea9c84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.928196] env[62952]: DEBUG nova.compute.provider_tree [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.930137] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.019731] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.066188] env[62952]: DEBUG nova.network.neutron [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.274975] env[62952]: DEBUG nova.compute.manager [req-9b32d3d0-3993-43eb-92fa-d2666efe13ce req-3cc5b65b-9af4-4ca4-8d01-5e820bc77a39 service nova] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Received event network-vif-deleted-aa82811b-2e65-4278-b2d4-40301d1e8b7d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.434151] env[62952]: DEBUG nova.scheduler.client.report [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.522576] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Releasing lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.522999] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.523207] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.523499] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c00d90a-d463-406a-92d4-444b5e222407 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.532027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9662a791-1070-403b-b277-53a1c9bec7d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.553152] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a224dd44-bda2-464f-b946-edbe167927b9 could not be found. [ 741.553385] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.553564] env[62952]: INFO nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 741.553802] env[62952]: DEBUG oslo.service.loopingcall [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.554035] env[62952]: DEBUG nova.compute.manager [-] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.554132] env[62952]: DEBUG nova.network.neutron [-] [instance: a224dd44-bda2-464f-b946-edbe167927b9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.569905] env[62952]: DEBUG nova.network.neutron [-] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.571307] env[62952]: INFO nova.compute.manager [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] [instance: 710a4cb2-a1f6-461e-ac26-d1e86cdd435d] Took 1.02 seconds to deallocate network for instance. [ 741.938788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.939225] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.942296] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.476s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.943710] env[62952]: INFO nova.compute.claims [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.077667] env[62952]: DEBUG nova.network.neutron [-] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.449557] env[62952]: DEBUG nova.compute.utils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.452725] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.452915] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.508399] env[62952]: DEBUG nova.policy [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '780753029d784f938f37ddc55fde80e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92a023e94514437dbed14b7e761d0a91', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.582406] env[62952]: INFO nova.compute.manager [-] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Took 1.03 seconds to deallocate network for instance. [ 742.586959] env[62952]: DEBUG nova.compute.claims [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 742.587125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.611695] env[62952]: INFO nova.scheduler.client.report [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Deleted allocations for instance 710a4cb2-a1f6-461e-ac26-d1e86cdd435d [ 742.815889] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Successfully created port: 36f66ae9-b16e-41fe-b029-370e6ec6b611 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.953485] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.124588] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2289a245-feb0-49e2-bb55-58ba6ed50850 tempest-ServerGroupTestJSON-1694021366 tempest-ServerGroupTestJSON-1694021366-project-member] Lock "710a4cb2-a1f6-461e-ac26-d1e86cdd435d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.763s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.373381] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d008aeaa-f920-46dd-9302-9cf8b67ebe1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.380293] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8e136d-b8ea-4c1d-b40a-a24ffa8771ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.411809] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a55f1b-ca84-45e8-ba6a-10f4fe25a0a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.419304] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97a4169-aa0a-4c1e-a4dd-e0fec86180a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.433860] env[62952]: DEBUG nova.compute.provider_tree [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.462180] env[62952]: INFO nova.virt.block_device [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Booting with volume b4c4ed52-bbfb-48f0-9ebe-4615188cbdd2 at /dev/sda [ 743.507526] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b4b547a-6e58-4c8a-b781-dfb21f8fb696 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.518771] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b427836b-5643-4d55-a3fd-591dd8bc90d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.541660] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a6b2a16-5d5c-4d6a-82d7-11467099869d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.549314] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ff1c79-0652-46cd-9f4c-eb6543e047af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.572147] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf7e13d-67c1-430e-8516-e9691ca04f35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.578860] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c752c6-195f-44fb-8e4b-7e3d52770305 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.593152] env[62952]: DEBUG nova.virt.block_device [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Updating existing volume attachment record: ec601d95-6074-4de7-8d98-e97820e25779 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 743.630171] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.937629] env[62952]: DEBUG nova.scheduler.client.report [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.051506] env[62952]: DEBUG nova.compute.manager [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Received event network-changed-36f66ae9-b16e-41fe-b029-370e6ec6b611 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.051778] env[62952]: DEBUG nova.compute.manager [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Refreshing instance network info cache due to event network-changed-36f66ae9-b16e-41fe-b029-370e6ec6b611. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.051943] env[62952]: DEBUG oslo_concurrency.lockutils [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] Acquiring lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.052070] env[62952]: DEBUG oslo_concurrency.lockutils [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] Acquired lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.052258] env[62952]: DEBUG nova.network.neutron [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Refreshing network info cache for port 36f66ae9-b16e-41fe-b029-370e6ec6b611 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.160602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.442627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.443013] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.445779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.447460] env[62952]: INFO nova.compute.claims [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.545823] env[62952]: ERROR nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 744.545823] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.545823] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.545823] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.545823] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.545823] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.545823] env[62952]: ERROR nova.compute.manager raise self.value [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.545823] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 744.545823] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.545823] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 744.546253] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.546253] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 744.546253] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 744.546253] env[62952]: ERROR nova.compute.manager [ 744.546253] env[62952]: Traceback (most recent call last): [ 744.546253] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 744.546253] env[62952]: listener.cb(fileno) [ 744.546253] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.546253] env[62952]: result = function(*args, **kwargs) [ 744.546253] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.546253] env[62952]: return func(*args, **kwargs) [ 744.546253] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.546253] env[62952]: raise e [ 744.546253] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.546253] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 744.546253] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.546253] env[62952]: created_port_ids = self._update_ports_for_instance( [ 744.546253] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.546253] env[62952]: with excutils.save_and_reraise_exception(): [ 744.546253] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.546253] env[62952]: self.force_reraise() [ 744.546253] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.546253] env[62952]: raise self.value [ 744.546253] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.546253] env[62952]: updated_port = self._update_port( [ 744.546253] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.546253] env[62952]: _ensure_no_port_binding_failure(port) [ 744.546253] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.546253] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 744.547736] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 744.547736] env[62952]: Removing descriptor: 19 [ 744.584138] env[62952]: DEBUG nova.network.neutron [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.712432] env[62952]: DEBUG nova.network.neutron [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.953821] env[62952]: DEBUG nova.compute.utils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.962998] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.963225] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.059075] env[62952]: DEBUG nova.policy [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6165653e44f147ee91e4a2edef90337e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd43742a77a784ed087a0105400e61435', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.215961] env[62952]: DEBUG oslo_concurrency.lockutils [req-0290980c-da5b-4364-812f-212b9cb57145 req-aabdaf0c-5d0d-42dd-85cf-9de31f01ae7c service nova] Releasing lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.355368] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Successfully created port: 72f3f1b8-4a28-4714-b5f5-ac0bcf514015 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.461052] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.715984] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.716682] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.716939] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.717128] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.717373] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.717543] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.717708] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.717981] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.722158] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.722422] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.722741] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.723053] env[62952]: DEBUG nova.virt.hardware [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.724119] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcf9f21-6036-4cb9-b2cf-fcf90a888635 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.734029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98281f6-e5dc-42dc-9c77-6fd26ca92ea3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.752150] env[62952]: ERROR nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Traceback (most recent call last): [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] yield resources [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self.driver.spawn(context, instance, image_meta, [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] vm_ref = self.build_virtual_machine(instance, [ 745.752150] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] for vif in network_info: [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] return self._sync_wrapper(fn, *args, **kwargs) [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self.wait() [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self[:] = self._gt.wait() [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] return self._exit_event.wait() [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 745.752502] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] current.throw(*self._exc) [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] result = function(*args, **kwargs) [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] return func(*args, **kwargs) [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] raise e [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] nwinfo = self.network_api.allocate_for_instance( [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] created_port_ids = self._update_ports_for_instance( [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] with excutils.save_and_reraise_exception(): [ 745.752825] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self.force_reraise() [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] raise self.value [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] updated_port = self._update_port( [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] _ensure_no_port_binding_failure(port) [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] raise exception.PortBindingFailed(port_id=port['id']) [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 745.753154] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] [ 745.753154] env[62952]: INFO nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Terminating instance [ 745.754868] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquiring lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.755065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquired lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.758504] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.923368] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3189b9c5-3dfe-454c-a217-d0fafdcf40fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.930575] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1673236-f41d-4f23-bf70-7b5fad9476e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.961510] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017dcfad-e078-4c91-81f6-ef9179e21bd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.972029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ebaa12-d3a7-469e-8d9c-833badbb90dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.985466] env[62952]: DEBUG nova.compute.provider_tree [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.087755] env[62952]: DEBUG nova.compute.manager [req-49f2a2f4-48d7-441f-9c38-0702965749a1 req-fc0f77ec-4b28-48b1-a962-9e0cbcaae522 service nova] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Received event network-vif-deleted-36f66ae9-b16e-41fe-b029-370e6ec6b611 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.284173] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.407765] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.446978] env[62952]: ERROR nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 746.446978] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.446978] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.446978] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.446978] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.446978] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.446978] env[62952]: ERROR nova.compute.manager raise self.value [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.446978] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.446978] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.446978] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.447586] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.447586] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.447586] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 746.447586] env[62952]: ERROR nova.compute.manager [ 746.447586] env[62952]: Traceback (most recent call last): [ 746.447586] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.447586] env[62952]: listener.cb(fileno) [ 746.447586] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.447586] env[62952]: result = function(*args, **kwargs) [ 746.447586] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.447586] env[62952]: return func(*args, **kwargs) [ 746.447586] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.447586] env[62952]: raise e [ 746.447586] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.447586] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 746.447586] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.447586] env[62952]: created_port_ids = self._update_ports_for_instance( [ 746.447586] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.447586] env[62952]: with excutils.save_and_reraise_exception(): [ 746.447586] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.447586] env[62952]: self.force_reraise() [ 746.447586] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.447586] env[62952]: raise self.value [ 746.447586] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.447586] env[62952]: updated_port = self._update_port( [ 746.447586] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.447586] env[62952]: _ensure_no_port_binding_failure(port) [ 746.447586] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.447586] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.448425] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 746.448425] env[62952]: Removing descriptor: 19 [ 746.478823] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.490940] env[62952]: DEBUG nova.scheduler.client.report [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.507334] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.507591] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.507745] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.507922] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.508077] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.508223] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.508425] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.508581] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.508741] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.508897] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.509085] env[62952]: DEBUG nova.virt.hardware [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.509930] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff2e9f6-3899-4bb5-8257-16dab1f848ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.520555] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5603d175-d883-45d3-a365-32b38f990672 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.534925] env[62952]: ERROR nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Traceback (most recent call last): [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] yield resources [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self.driver.spawn(context, instance, image_meta, [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] vm_ref = self.build_virtual_machine(instance, [ 746.534925] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] for vif in network_info: [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] return self._sync_wrapper(fn, *args, **kwargs) [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self.wait() [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self[:] = self._gt.wait() [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] return self._exit_event.wait() [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.535286] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] current.throw(*self._exc) [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] result = function(*args, **kwargs) [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] return func(*args, **kwargs) [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] raise e [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] nwinfo = self.network_api.allocate_for_instance( [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] created_port_ids = self._update_ports_for_instance( [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] with excutils.save_and_reraise_exception(): [ 746.535662] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self.force_reraise() [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] raise self.value [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] updated_port = self._update_port( [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] _ensure_no_port_binding_failure(port) [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] raise exception.PortBindingFailed(port_id=port['id']) [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 746.536046] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] [ 746.536046] env[62952]: INFO nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Terminating instance [ 746.537377] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.537672] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquired lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.537738] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.909536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Releasing lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.910146] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 746.910617] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8c20e21c-1b86-440a-a196-e1a9465fdcb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.918853] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3832ef-5a53-4e49-988f-a52a988da4e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.938916] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 869c5e7f-b937-4add-ace2-7d9b178c2acc could not be found. [ 746.939122] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 746.939368] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1256e104-201a-4ad8-ac92-ebf920b15b32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.946158] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8fc91b-c154-4d89-9efe-f93511d4cfd8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.967048] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 869c5e7f-b937-4add-ace2-7d9b178c2acc could not be found. [ 746.967211] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 746.967386] env[62952]: INFO nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Took 0.06 seconds to destroy the instance on the hypervisor. [ 746.967612] env[62952]: DEBUG oslo.service.loopingcall [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.967816] env[62952]: DEBUG nova.compute.manager [-] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.967916] env[62952]: DEBUG nova.network.neutron [-] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.981931] env[62952]: DEBUG nova.network.neutron [-] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.996284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.996730] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.999721] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.544s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.055538] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.123806] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.486571] env[62952]: DEBUG nova.network.neutron [-] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.506308] env[62952]: DEBUG nova.compute.utils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.508441] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.508441] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.557570] env[62952]: DEBUG nova.policy [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb541b4abd9426ea96edcd066c69be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d1f373267464f14ae2c833151821973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.626603] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Releasing lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.627212] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.627383] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.628210] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0280fe38-3be5-479c-847e-e4b3d9c5a1c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.637130] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339a159a-3eb2-466c-98b7-4bbc0e222bce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.665187] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac386ddb-c812-4689-b9b4-b98d26ad1831 could not be found. [ 747.665479] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.665613] env[62952]: INFO nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Took 0.04 seconds to destroy the instance on the hypervisor. [ 747.665886] env[62952]: DEBUG oslo.service.loopingcall [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.668959] env[62952]: DEBUG nova.compute.manager [-] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.669085] env[62952]: DEBUG nova.network.neutron [-] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.698214] env[62952]: DEBUG nova.network.neutron [-] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.939403] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db062dea-d494-4243-9c3d-c75e83ff268e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.947865] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6567164f-5267-44c9-8e88-a71c31f262c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.982267] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab34dcf-be48-43b6-ba81-07f3925a39aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.990082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c039c39d-3574-490d-b80f-0b70e3b253b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.994688] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Successfully created port: e00b15fd-a809-45b2-a3ff-92c61ade6abe {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.997032] env[62952]: INFO nova.compute.manager [-] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Took 1.03 seconds to deallocate network for instance. [ 748.011112] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.015695] env[62952]: DEBUG nova.compute.provider_tree [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.113111] env[62952]: DEBUG nova.compute.manager [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Received event network-changed-72f3f1b8-4a28-4714-b5f5-ac0bcf514015 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.113111] env[62952]: DEBUG nova.compute.manager [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Refreshing instance network info cache due to event network-changed-72f3f1b8-4a28-4714-b5f5-ac0bcf514015. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.113652] env[62952]: DEBUG oslo_concurrency.lockutils [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] Acquiring lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.113652] env[62952]: DEBUG oslo_concurrency.lockutils [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] Acquired lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.113652] env[62952]: DEBUG nova.network.neutron [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Refreshing network info cache for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 748.200559] env[62952]: DEBUG nova.network.neutron [-] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.523542] env[62952]: DEBUG nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.580993] env[62952]: INFO nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Took 0.58 seconds to detach 1 volumes for instance. [ 748.583221] env[62952]: DEBUG nova.compute.claims [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 748.583347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.643638] env[62952]: DEBUG nova.network.neutron [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.702757] env[62952]: INFO nova.compute.manager [-] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Took 1.03 seconds to deallocate network for instance. [ 748.710847] env[62952]: DEBUG nova.compute.claims [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 748.710847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.751783] env[62952]: DEBUG nova.network.neutron [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.990343] env[62952]: ERROR nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 748.990343] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.990343] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.990343] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.990343] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.990343] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.990343] env[62952]: ERROR nova.compute.manager raise self.value [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.990343] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 748.990343] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.990343] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 748.990892] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.990892] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 748.990892] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 748.990892] env[62952]: ERROR nova.compute.manager [ 748.990892] env[62952]: Traceback (most recent call last): [ 748.990892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 748.990892] env[62952]: listener.cb(fileno) [ 748.990892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.990892] env[62952]: result = function(*args, **kwargs) [ 748.990892] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.990892] env[62952]: return func(*args, **kwargs) [ 748.990892] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.990892] env[62952]: raise e [ 748.990892] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.990892] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 748.990892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.990892] env[62952]: created_port_ids = self._update_ports_for_instance( [ 748.990892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.990892] env[62952]: with excutils.save_and_reraise_exception(): [ 748.990892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.990892] env[62952]: self.force_reraise() [ 748.990892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.990892] env[62952]: raise self.value [ 748.990892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.990892] env[62952]: updated_port = self._update_port( [ 748.990892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.990892] env[62952]: _ensure_no_port_binding_failure(port) [ 748.990892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.990892] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 748.991734] env[62952]: nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 748.991734] env[62952]: Removing descriptor: 19 [ 749.032701] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.035300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.035943] env[62952]: ERROR nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Traceback (most recent call last): [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self.driver.spawn(context, instance, image_meta, [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] vm_ref = self.build_virtual_machine(instance, [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.035943] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] for vif in network_info: [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return self._sync_wrapper(fn, *args, **kwargs) [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self.wait() [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self[:] = self._gt.wait() [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return self._exit_event.wait() [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] result = hub.switch() [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.036269] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return self.greenlet.switch() [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] result = function(*args, **kwargs) [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] return func(*args, **kwargs) [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] raise e [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] nwinfo = self.network_api.allocate_for_instance( [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] created_port_ids = self._update_ports_for_instance( [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] with excutils.save_and_reraise_exception(): [ 749.036591] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] self.force_reraise() [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] raise self.value [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] updated_port = self._update_port( [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] _ensure_no_port_binding_failure(port) [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] raise exception.PortBindingFailed(port_id=port['id']) [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] nova.exception.PortBindingFailed: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. [ 749.036898] env[62952]: ERROR nova.compute.manager [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] [ 749.037174] env[62952]: DEBUG nova.compute.utils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.039175] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Build of instance de2daf8c-73c1-44fe-bb01-2b54f1f1c419 was re-scheduled: Binding failed for port d208e472-8506-4ab4-8e8a-a63b5e63a3d9, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.039636] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.039898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.040091] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquired lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.040291] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.041459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.916s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.043505] env[62952]: INFO nova.compute.claims [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.065934] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.066193] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.066350] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.066528] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.066675] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.066821] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.067034] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.067198] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.067363] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.067521] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.067819] env[62952]: DEBUG nova.virt.hardware [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.068770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2872c28e-e8b7-45d6-a2f7-4d35004bd09e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.076908] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84aa7c3-fbc7-4338-b3d7-4aa6f158ff71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.092210] env[62952]: ERROR nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Traceback (most recent call last): [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] yield resources [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self.driver.spawn(context, instance, image_meta, [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] vm_ref = self.build_virtual_machine(instance, [ 749.092210] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] for vif in network_info: [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] return self._sync_wrapper(fn, *args, **kwargs) [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self.wait() [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self[:] = self._gt.wait() [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] return self._exit_event.wait() [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.092689] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] current.throw(*self._exc) [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] result = function(*args, **kwargs) [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] return func(*args, **kwargs) [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] raise e [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] nwinfo = self.network_api.allocate_for_instance( [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] created_port_ids = self._update_ports_for_instance( [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] with excutils.save_and_reraise_exception(): [ 749.093103] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self.force_reraise() [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] raise self.value [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] updated_port = self._update_port( [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] _ensure_no_port_binding_failure(port) [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] raise exception.PortBindingFailed(port_id=port['id']) [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 749.093486] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] [ 749.093486] env[62952]: INFO nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Terminating instance [ 749.094318] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.094478] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.094643] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.254325] env[62952]: DEBUG oslo_concurrency.lockutils [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] Releasing lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.254641] env[62952]: DEBUG nova.compute.manager [req-c7624939-e247-4c73-93ca-3f8dc5b54667 req-24106361-2fc0-4359-956c-df231f25981d service nova] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Received event network-vif-deleted-72f3f1b8-4a28-4714-b5f5-ac0bcf514015 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.563483] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.625652] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.840917] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.907323] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.146823] env[62952]: DEBUG nova.compute.manager [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Received event network-changed-e00b15fd-a809-45b2-a3ff-92c61ade6abe {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.147037] env[62952]: DEBUG nova.compute.manager [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Refreshing instance network info cache due to event network-changed-e00b15fd-a809-45b2-a3ff-92c61ade6abe. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.147231] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] Acquiring lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.344905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Releasing lock "refresh_cache-de2daf8c-73c1-44fe-bb01-2b54f1f1c419" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.345177] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.345331] env[62952]: DEBUG nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.345498] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.359660] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.368114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b354ba-adec-4484-9798-82f1ef762b90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.375907] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4fe864-0dae-4bc2-9c94-ccce90c26c55 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.405501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72c6c72-d8d7-452d-8e32-e89d0594f94a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.411607] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.411985] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 750.412192] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.412504] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] Acquired lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.412715] env[62952]: DEBUG nova.network.neutron [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Refreshing network info cache for port e00b15fd-a809-45b2-a3ff-92c61ade6abe {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.413807] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ac95594-0300-46f5-9974-69d92c31cb02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.416753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c63db7-571c-4ee5-80e0-6062179b34ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.431386] env[62952]: DEBUG nova.compute.provider_tree [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.435987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecc4d88-ae12-47d9-9141-66a84b395220 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.457083] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f4ec1e94-1778-4296-989d-782700c6faf8 could not be found. [ 750.457292] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 750.457471] env[62952]: INFO nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Took 0.05 seconds to destroy the instance on the hypervisor. [ 750.457705] env[62952]: DEBUG oslo.service.loopingcall [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.457903] env[62952]: DEBUG nova.compute.manager [-] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.458043] env[62952]: DEBUG nova.network.neutron [-] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.472072] env[62952]: DEBUG nova.network.neutron [-] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.863133] env[62952]: DEBUG nova.network.neutron [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.934911] env[62952]: DEBUG nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.943561] env[62952]: DEBUG nova.network.neutron [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.975613] env[62952]: DEBUG nova.network.neutron [-] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.023706] env[62952]: DEBUG nova.network.neutron [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.366264] env[62952]: INFO nova.compute.manager [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: de2daf8c-73c1-44fe-bb01-2b54f1f1c419] Took 1.02 seconds to deallocate network for instance. [ 751.439652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.440187] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.442739] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.336s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.444103] env[62952]: INFO nova.compute.claims [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.476725] env[62952]: INFO nova.compute.manager [-] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Took 1.02 seconds to deallocate network for instance. [ 751.478815] env[62952]: DEBUG nova.compute.claims [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 751.478988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.526853] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] Releasing lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.527131] env[62952]: DEBUG nova.compute.manager [req-9a3e3dea-888f-4da7-bc7a-4ece964485ef req-0c1d3c72-f23b-47a0-abfd-44ddbe6b62c8 service nova] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Received event network-vif-deleted-e00b15fd-a809-45b2-a3ff-92c61ade6abe {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.949712] env[62952]: DEBUG nova.compute.utils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.952449] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.953894] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.003539] env[62952]: DEBUG nova.policy [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74e0e37130134da6bb3a2c4c33d2efa8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd14bf89377d4e37b19e787cc5234033', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.297415] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Successfully created port: e4596600-6a36-4ccc-a18f-b514c54f1f33 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.398020] env[62952]: INFO nova.scheduler.client.report [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Deleted allocations for instance de2daf8c-73c1-44fe-bb01-2b54f1f1c419 [ 752.453394] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.815245] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811f1db3-fd84-40cb-bb75-a243f5ec632b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.822919] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcdd7b5-61a8-4f4c-bfa1-4b2f57ebba5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.852411] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9013996d-cb53-4089-b2d2-8c90f2b7473d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.859372] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95163170-c6a8-4d2b-90d4-a01854f696b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.872172] env[62952]: DEBUG nova.compute.provider_tree [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.910500] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdc4ff16-a981-4c06-bd50-d4e3a92f46f7 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "de2daf8c-73c1-44fe-bb01-2b54f1f1c419" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.027s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.019387] env[62952]: DEBUG nova.compute.manager [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Received event network-changed-e4596600-6a36-4ccc-a18f-b514c54f1f33 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.019577] env[62952]: DEBUG nova.compute.manager [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Refreshing instance network info cache due to event network-changed-e4596600-6a36-4ccc-a18f-b514c54f1f33. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.019827] env[62952]: DEBUG oslo_concurrency.lockutils [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] Acquiring lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.019971] env[62952]: DEBUG oslo_concurrency.lockutils [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] Acquired lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.020165] env[62952]: DEBUG nova.network.neutron [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Refreshing network info cache for port e4596600-6a36-4ccc-a18f-b514c54f1f33 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.217912] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 753.217912] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.217912] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.217912] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.217912] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.217912] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.217912] env[62952]: ERROR nova.compute.manager raise self.value [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.217912] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.217912] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.217912] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.218425] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.218425] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.218425] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 753.218425] env[62952]: ERROR nova.compute.manager [ 753.218425] env[62952]: Traceback (most recent call last): [ 753.218425] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.218425] env[62952]: listener.cb(fileno) [ 753.218425] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.218425] env[62952]: result = function(*args, **kwargs) [ 753.218425] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.218425] env[62952]: return func(*args, **kwargs) [ 753.218425] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.218425] env[62952]: raise e [ 753.218425] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.218425] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 753.218425] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.218425] env[62952]: created_port_ids = self._update_ports_for_instance( [ 753.218425] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.218425] env[62952]: with excutils.save_and_reraise_exception(): [ 753.218425] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.218425] env[62952]: self.force_reraise() [ 753.218425] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.218425] env[62952]: raise self.value [ 753.218425] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.218425] env[62952]: updated_port = self._update_port( [ 753.218425] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.218425] env[62952]: _ensure_no_port_binding_failure(port) [ 753.218425] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.218425] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.219248] env[62952]: nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 753.219248] env[62952]: Removing descriptor: 15 [ 753.376060] env[62952]: DEBUG nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.413502] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.469568] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.494445] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.494683] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.494836] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.495020] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.495161] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.495301] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.495495] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.495647] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.495806] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.496038] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.498051] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.498149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1874567-a603-4c44-abb9-3e01aff32187 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.505819] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3eda1d1-42b5-4783-a632-62bc82fb2e73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.527183] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Traceback (most recent call last): [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] yield resources [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self.driver.spawn(context, instance, image_meta, [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] vm_ref = self.build_virtual_machine(instance, [ 753.527183] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] for vif in network_info: [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] return self._sync_wrapper(fn, *args, **kwargs) [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self.wait() [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self[:] = self._gt.wait() [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] return self._exit_event.wait() [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.527745] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] current.throw(*self._exc) [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] result = function(*args, **kwargs) [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] return func(*args, **kwargs) [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] raise e [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] nwinfo = self.network_api.allocate_for_instance( [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] created_port_ids = self._update_ports_for_instance( [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] with excutils.save_and_reraise_exception(): [ 753.529277] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self.force_reraise() [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] raise self.value [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] updated_port = self._update_port( [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] _ensure_no_port_binding_failure(port) [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] raise exception.PortBindingFailed(port_id=port['id']) [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 753.529724] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] [ 753.529724] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Terminating instance [ 753.530214] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.542130] env[62952]: DEBUG nova.network.neutron [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.636730] env[62952]: DEBUG nova.network.neutron [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.883877] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.884446] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 753.888809] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.782s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.888988] env[62952]: INFO nova.compute.claims [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.940283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.139809] env[62952]: DEBUG oslo_concurrency.lockutils [req-7b03e448-c5ee-44ea-a888-ba7c1d0011df req-a7883b3c-de57-4fa6-bf50-60709ec81d23 service nova] Releasing lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.140020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquired lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.140210] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.393042] env[62952]: DEBUG nova.compute.utils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.397525] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 754.397694] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.463729] env[62952]: DEBUG nova.policy [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74e0e37130134da6bb3a2c4c33d2efa8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd14bf89377d4e37b19e787cc5234033', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.672617] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.799168] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.891981] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Successfully created port: b1f7df5e-c0d9-494e-8185-b8f132bf3a44 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.896567] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.049221] env[62952]: DEBUG nova.compute.manager [req-70bb8e7a-ecaa-45f4-a692-383981174f9a req-9acfa94a-1de9-4339-8b71-31c9ba7b56a8 service nova] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Received event network-vif-deleted-e4596600-6a36-4ccc-a18f-b514c54f1f33 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.302058] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Releasing lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.302465] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.302653] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.303172] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbb03bf7-d39d-4bba-a766-489aaef29974 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.312915] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a216f1f6-ca64-414e-b505-fc21ef27e264 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.338559] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71 could not be found. [ 755.338960] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.339196] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Took 0.04 seconds to destroy the instance on the hypervisor. [ 755.339510] env[62952]: DEBUG oslo.service.loopingcall [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.339787] env[62952]: DEBUG nova.compute.manager [-] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.339881] env[62952]: DEBUG nova.network.neutron [-] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.343113] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3f6bac-ba3a-4fba-8abe-ce03e7184a92 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.349800] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f0a3a5-ed32-4593-9374-6033efc5b091 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.383845] env[62952]: DEBUG nova.network.neutron [-] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.385629] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876db753-d806-461c-80cf-941d14714002 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.395067] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628dc40d-b27b-449e-a57d-5445290f0025 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.411657] env[62952]: DEBUG nova.compute.provider_tree [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.890010] env[62952]: DEBUG nova.network.neutron [-] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.914470] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 755.917285] env[62952]: DEBUG nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.942565] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.942811] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.942988] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.943192] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.943384] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.943484] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.943685] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.943846] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.944096] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.944289] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.944469] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.946150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d5199e-1b14-495c-87a6-dc8035ccea2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.954276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91da5f7-8a96-4781-a210-698f717bc22a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.970580] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 755.970580] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.970580] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.970580] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.970580] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.970580] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.970580] env[62952]: ERROR nova.compute.manager raise self.value [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.970580] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 755.970580] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.970580] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 755.971070] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.971070] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 755.971070] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 755.971070] env[62952]: ERROR nova.compute.manager [ 755.971070] env[62952]: Traceback (most recent call last): [ 755.971070] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 755.971070] env[62952]: listener.cb(fileno) [ 755.971070] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.971070] env[62952]: result = function(*args, **kwargs) [ 755.971070] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 755.971070] env[62952]: return func(*args, **kwargs) [ 755.971070] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.971070] env[62952]: raise e [ 755.971070] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.971070] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 755.971070] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.971070] env[62952]: created_port_ids = self._update_ports_for_instance( [ 755.971070] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.971070] env[62952]: with excutils.save_and_reraise_exception(): [ 755.971070] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.971070] env[62952]: self.force_reraise() [ 755.971070] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.971070] env[62952]: raise self.value [ 755.971070] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.971070] env[62952]: updated_port = self._update_port( [ 755.971070] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.971070] env[62952]: _ensure_no_port_binding_failure(port) [ 755.971070] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.971070] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 755.971896] env[62952]: nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 755.971896] env[62952]: Removing descriptor: 15 [ 755.971896] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Traceback (most recent call last): [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] yield resources [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self.driver.spawn(context, instance, image_meta, [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.971896] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] vm_ref = self.build_virtual_machine(instance, [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] for vif in network_info: [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return self._sync_wrapper(fn, *args, **kwargs) [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self.wait() [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self[:] = self._gt.wait() [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return self._exit_event.wait() [ 755.972256] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] result = hub.switch() [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return self.greenlet.switch() [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] result = function(*args, **kwargs) [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return func(*args, **kwargs) [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] raise e [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] nwinfo = self.network_api.allocate_for_instance( [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.972601] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] created_port_ids = self._update_ports_for_instance( [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] with excutils.save_and_reraise_exception(): [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self.force_reraise() [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] raise self.value [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] updated_port = self._update_port( [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] _ensure_no_port_binding_failure(port) [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.973031] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] raise exception.PortBindingFailed(port_id=port['id']) [ 755.973362] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 755.973362] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] [ 755.973362] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Terminating instance [ 755.973497] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.973653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquired lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.973813] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.223706] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "39948315-6438-48f1-883b-b376e0786650" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.223933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "39948315-6438-48f1-883b-b376e0786650" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.394629] env[62952]: INFO nova.compute.manager [-] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Took 1.05 seconds to deallocate network for instance. [ 756.396014] env[62952]: DEBUG nova.compute.claims [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 756.396210] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.423147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.423640] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.428696] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.962s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.493126] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.583552] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.928153] env[62952]: DEBUG nova.compute.utils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.929558] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.929734] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.985373] env[62952]: DEBUG nova.policy [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74e0e37130134da6bb3a2c4c33d2efa8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd14bf89377d4e37b19e787cc5234033', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.075896] env[62952]: DEBUG nova.compute.manager [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Received event network-changed-b1f7df5e-c0d9-494e-8185-b8f132bf3a44 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.076116] env[62952]: DEBUG nova.compute.manager [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Refreshing instance network info cache due to event network-changed-b1f7df5e-c0d9-494e-8185-b8f132bf3a44. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.076281] env[62952]: DEBUG oslo_concurrency.lockutils [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] Acquiring lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.085883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Releasing lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.086294] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.086482] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.087012] env[62952]: DEBUG oslo_concurrency.lockutils [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] Acquired lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.087188] env[62952]: DEBUG nova.network.neutron [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Refreshing network info cache for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.088259] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1bf8b30e-8f21-4e7a-90c2-b561600ab27c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.096834] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8257fca-9812-4c35-aa2e-e89194a392b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.121865] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a825d324-0b26-4051-bdfa-4a4246f6d4e9 could not be found. [ 757.122103] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.122288] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 757.122530] env[62952]: DEBUG oslo.service.loopingcall [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.122744] env[62952]: DEBUG nova.compute.manager [-] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.123127] env[62952]: DEBUG nova.network.neutron [-] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.146302] env[62952]: DEBUG nova.network.neutron [-] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.264640] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Successfully created port: 1463b38b-65ad-49e0-a75d-3dd3e86e3438 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.303791] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bc4c1f-3ae2-4a95-88d9-1d0a4faf5e88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.313191] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524173d5-6235-450c-8436-71333ead982b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.342861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48578657-a22a-4a4a-99a2-689eabed7794 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.351070] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be46a5c-7a11-4b91-b4b2-d7b594ce97bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.365114] env[62952]: DEBUG nova.compute.provider_tree [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.433402] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.607188] env[62952]: DEBUG nova.network.neutron [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.649395] env[62952]: DEBUG nova.network.neutron [-] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.740206] env[62952]: DEBUG nova.network.neutron [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.868021] env[62952]: DEBUG nova.scheduler.client.report [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.152089] env[62952]: INFO nova.compute.manager [-] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Took 1.03 seconds to deallocate network for instance. [ 758.154449] env[62952]: DEBUG nova.compute.claims [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 758.154680] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.242927] env[62952]: DEBUG oslo_concurrency.lockutils [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] Releasing lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.243301] env[62952]: DEBUG nova.compute.manager [req-b4fedada-c096-4913-b1b9-c76e5b477e03 req-bbe4552a-a5e5-4d8f-9dd8-023505f92228 service nova] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Received event network-vif-deleted-b1f7df5e-c0d9-494e-8185-b8f132bf3a44 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.373218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.373872] env[62952]: ERROR nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Traceback (most recent call last): [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self.driver.spawn(context, instance, image_meta, [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] vm_ref = self.build_virtual_machine(instance, [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.373872] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] for vif in network_info: [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return self._sync_wrapper(fn, *args, **kwargs) [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self.wait() [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self[:] = self._gt.wait() [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return self._exit_event.wait() [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] result = hub.switch() [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.374268] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return self.greenlet.switch() [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] result = function(*args, **kwargs) [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] return func(*args, **kwargs) [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] raise e [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] nwinfo = self.network_api.allocate_for_instance( [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] created_port_ids = self._update_ports_for_instance( [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] with excutils.save_and_reraise_exception(): [ 758.374639] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] self.force_reraise() [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] raise self.value [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] updated_port = self._update_port( [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] _ensure_no_port_binding_failure(port) [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] raise exception.PortBindingFailed(port_id=port['id']) [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] nova.exception.PortBindingFailed: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. [ 758.375018] env[62952]: ERROR nova.compute.manager [instance: 244c71ad-130d-44ad-9998-3f5c51518482] [ 758.375343] env[62952]: DEBUG nova.compute.utils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 758.375836] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.789s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.378610] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Build of instance 244c71ad-130d-44ad-9998-3f5c51518482 was re-scheduled: Binding failed for port fadeb616-d710-4ea9-897e-1f887a92b3f2, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 758.379041] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 758.379266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.379412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquired lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.379568] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.411855] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 758.411855] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.411855] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.411855] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.411855] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.411855] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.411855] env[62952]: ERROR nova.compute.manager raise self.value [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.411855] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.411855] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.411855] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.412993] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.412993] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.412993] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 758.412993] env[62952]: ERROR nova.compute.manager [ 758.412993] env[62952]: Traceback (most recent call last): [ 758.412993] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.412993] env[62952]: listener.cb(fileno) [ 758.412993] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.412993] env[62952]: result = function(*args, **kwargs) [ 758.412993] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.412993] env[62952]: return func(*args, **kwargs) [ 758.412993] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.412993] env[62952]: raise e [ 758.412993] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.412993] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 758.412993] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.412993] env[62952]: created_port_ids = self._update_ports_for_instance( [ 758.412993] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.412993] env[62952]: with excutils.save_and_reraise_exception(): [ 758.412993] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.412993] env[62952]: self.force_reraise() [ 758.412993] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.412993] env[62952]: raise self.value [ 758.412993] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.412993] env[62952]: updated_port = self._update_port( [ 758.412993] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.412993] env[62952]: _ensure_no_port_binding_failure(port) [ 758.412993] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.412993] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.413915] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 758.413915] env[62952]: Removing descriptor: 15 [ 758.441615] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.472511] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.472752] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.472905] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.473123] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.473283] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.473429] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.473629] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.473784] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.473944] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.475882] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.475882] env[62952]: DEBUG nova.virt.hardware [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.476336] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d541ec-2658-44a3-a72d-8cc1a7dbf299 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.484281] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320584b3-2c72-4633-b3d1-54249ce0d870 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.500371] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Traceback (most recent call last): [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] yield resources [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self.driver.spawn(context, instance, image_meta, [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] vm_ref = self.build_virtual_machine(instance, [ 758.500371] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] for vif in network_info: [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] return self._sync_wrapper(fn, *args, **kwargs) [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self.wait() [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self[:] = self._gt.wait() [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] return self._exit_event.wait() [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.500765] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] current.throw(*self._exc) [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] result = function(*args, **kwargs) [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] return func(*args, **kwargs) [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] raise e [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] nwinfo = self.network_api.allocate_for_instance( [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] created_port_ids = self._update_ports_for_instance( [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] with excutils.save_and_reraise_exception(): [ 758.501148] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self.force_reraise() [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] raise self.value [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] updated_port = self._update_port( [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] _ensure_no_port_binding_failure(port) [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] raise exception.PortBindingFailed(port_id=port['id']) [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 758.501534] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] [ 758.501534] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Terminating instance [ 758.505313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.505477] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquired lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.505646] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.901421] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.968972] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.024747] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.109865] env[62952]: DEBUG nova.compute.manager [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Received event network-changed-1463b38b-65ad-49e0-a75d-3dd3e86e3438 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.110042] env[62952]: DEBUG nova.compute.manager [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Refreshing instance network info cache due to event network-changed-1463b38b-65ad-49e0-a75d-3dd3e86e3438. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.110239] env[62952]: DEBUG oslo_concurrency.lockutils [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] Acquiring lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.111071] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.209040] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a531122-594c-4032-b00e-fd0a28e1e847 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.215673] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a2095a-6476-41f1-ab6b-2818d7e4b75d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.247546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f05cdf-3a0c-41cf-b59f-47f873f470be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.254517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25890775-2b3a-4638-8819-af9362c6310d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.267706] env[62952]: DEBUG nova.compute.provider_tree [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.471695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Releasing lock "refresh_cache-244c71ad-130d-44ad-9998-3f5c51518482" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.472511] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 759.472511] env[62952]: DEBUG nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.472511] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.492290] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.613968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Releasing lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.614482] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.614689] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.615064] env[62952]: DEBUG oslo_concurrency.lockutils [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] Acquired lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.615254] env[62952]: DEBUG nova.network.neutron [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Refreshing network info cache for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 759.616350] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f8d3cae-3811-461a-9ba0-e77ed3380abc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.626215] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b49cdc-e3f9-4064-ba21-e5b4dcd2693d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.647535] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec could not be found. [ 759.647732] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.647919] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.648174] env[62952]: DEBUG oslo.service.loopingcall [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.648416] env[62952]: DEBUG nova.compute.manager [-] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.648475] env[62952]: DEBUG nova.network.neutron [-] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.669671] env[62952]: DEBUG nova.network.neutron [-] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.771054] env[62952]: DEBUG nova.scheduler.client.report [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.996050] env[62952]: DEBUG nova.network.neutron [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.134744] env[62952]: DEBUG nova.network.neutron [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.172451] env[62952]: DEBUG nova.network.neutron [-] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.205909] env[62952]: DEBUG nova.network.neutron [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.276049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.276918] env[62952]: ERROR nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] Traceback (most recent call last): [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self.driver.spawn(context, instance, image_meta, [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] vm_ref = self.build_virtual_machine(instance, [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.276918] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] for vif in network_info: [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] return self._sync_wrapper(fn, *args, **kwargs) [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self.wait() [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self[:] = self._gt.wait() [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] return self._exit_event.wait() [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] current.throw(*self._exc) [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.277408] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] result = function(*args, **kwargs) [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] return func(*args, **kwargs) [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] raise e [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] nwinfo = self.network_api.allocate_for_instance( [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] created_port_ids = self._update_ports_for_instance( [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] with excutils.save_and_reraise_exception(): [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] self.force_reraise() [ 760.277792] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] raise self.value [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] updated_port = self._update_port( [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] _ensure_no_port_binding_failure(port) [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] raise exception.PortBindingFailed(port_id=port['id']) [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] nova.exception.PortBindingFailed: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. [ 760.278245] env[62952]: ERROR nova.compute.manager [instance: a224dd44-bda2-464f-b946-edbe167927b9] [ 760.278245] env[62952]: DEBUG nova.compute.utils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.279868] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.119s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.282143] env[62952]: INFO nova.compute.claims [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.287289] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Build of instance a224dd44-bda2-464f-b946-edbe167927b9 was re-scheduled: Binding failed for port aa82811b-2e65-4278-b2d4-40301d1e8b7d, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 760.287289] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 760.287289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquiring lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.287289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Acquired lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.287536] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.498704] env[62952]: INFO nova.compute.manager [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: 244c71ad-130d-44ad-9998-3f5c51518482] Took 1.03 seconds to deallocate network for instance. [ 760.675045] env[62952]: INFO nova.compute.manager [-] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Took 1.03 seconds to deallocate network for instance. [ 760.676714] env[62952]: DEBUG nova.compute.claims [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.676886] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.708526] env[62952]: DEBUG oslo_concurrency.lockutils [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] Releasing lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.708766] env[62952]: DEBUG nova.compute.manager [req-97c75051-f5a2-4835-957d-38ef2390967d req-71c02ca9-5eb5-4456-bf82-ca4b0eeb7b2c service nova] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Received event network-vif-deleted-1463b38b-65ad-49e0-a75d-3dd3e86e3438 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.805732] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.874326] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.376550] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Releasing lock "refresh_cache-a224dd44-bda2-464f-b946-edbe167927b9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.377030] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.377271] env[62952]: DEBUG nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.377477] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.395412] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.528144] env[62952]: INFO nova.scheduler.client.report [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Deleted allocations for instance 244c71ad-130d-44ad-9998-3f5c51518482 [ 761.598860] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb29c287-3a63-4d3c-be30-fe93e79a2cca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.606507] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff5b66b-2ac0-404b-af86-4c115b3c55b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.636320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b578e287-a95b-4c97-970e-20b31a46bd5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.643719] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e356c517-19b1-4109-87b0-111ba022fd77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.656474] env[62952]: DEBUG nova.compute.provider_tree [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.897953] env[62952]: DEBUG nova.network.neutron [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.039152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-477d4fea-000d-4f3c-b021-45ab8d9f1c4a tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "244c71ad-130d-44ad-9998-3f5c51518482" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.653s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.161106] env[62952]: DEBUG nova.scheduler.client.report [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.400998] env[62952]: INFO nova.compute.manager [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] [instance: a224dd44-bda2-464f-b946-edbe167927b9] Took 1.02 seconds to deallocate network for instance. [ 762.542310] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.666727] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.666727] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.669667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.086s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.064666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.175715] env[62952]: DEBUG nova.compute.utils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.176534] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.176718] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.253596] env[62952]: DEBUG nova.policy [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '181708cc7de14c83ab99dc9017699f19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc2d66dc7eca4dfc975713f73a7ce13d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.429591] env[62952]: INFO nova.scheduler.client.report [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Deleted allocations for instance a224dd44-bda2-464f-b946-edbe167927b9 [ 763.576949] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Successfully created port: a6e81871-2452-4c69-bfe2-64804255718e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.580661] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe0c5b7-4cd7-4c6e-9794-fcaab617228a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.588659] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691b5d99-6e1e-4e7d-b440-beaa5f27ef24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.619388] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a790fd-f801-48a0-ae3c-5e6cd17eb59f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.627718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6742be-174e-4a80-a71a-944792c3beeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.644938] env[62952]: DEBUG nova.compute.provider_tree [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.687325] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.939898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea51304c-4917-4e09-9d66-7abf0a053d0a tempest-ServerMetadataNegativeTestJSON-401377393 tempest-ServerMetadataNegativeTestJSON-401377393-project-member] Lock "a224dd44-bda2-464f-b946-edbe167927b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.194s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.146818] env[62952]: DEBUG nova.scheduler.client.report [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.444650] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.481577] env[62952]: DEBUG nova.compute.manager [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Received event network-changed-a6e81871-2452-4c69-bfe2-64804255718e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.481769] env[62952]: DEBUG nova.compute.manager [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Refreshing instance network info cache due to event network-changed-a6e81871-2452-4c69-bfe2-64804255718e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.481981] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] Acquiring lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.482136] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] Acquired lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.482335] env[62952]: DEBUG nova.network.neutron [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Refreshing network info cache for port a6e81871-2452-4c69-bfe2-64804255718e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 764.615638] env[62952]: ERROR nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 764.615638] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.615638] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.615638] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.615638] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.615638] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.615638] env[62952]: ERROR nova.compute.manager raise self.value [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.615638] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.615638] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.615638] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.616144] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.616144] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.616144] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 764.616144] env[62952]: ERROR nova.compute.manager [ 764.616144] env[62952]: Traceback (most recent call last): [ 764.616144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.616144] env[62952]: listener.cb(fileno) [ 764.616144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.616144] env[62952]: result = function(*args, **kwargs) [ 764.616144] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.616144] env[62952]: return func(*args, **kwargs) [ 764.616144] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.616144] env[62952]: raise e [ 764.616144] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.616144] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 764.616144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.616144] env[62952]: created_port_ids = self._update_ports_for_instance( [ 764.616144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.616144] env[62952]: with excutils.save_and_reraise_exception(): [ 764.616144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.616144] env[62952]: self.force_reraise() [ 764.616144] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.616144] env[62952]: raise self.value [ 764.616144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.616144] env[62952]: updated_port = self._update_port( [ 764.616144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.616144] env[62952]: _ensure_no_port_binding_failure(port) [ 764.616144] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.616144] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.617076] env[62952]: nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 764.617076] env[62952]: Removing descriptor: 15 [ 764.655071] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.655071] env[62952]: ERROR nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Traceback (most recent call last): [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self.driver.spawn(context, instance, image_meta, [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.655071] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] vm_ref = self.build_virtual_machine(instance, [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] for vif in network_info: [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] return self._sync_wrapper(fn, *args, **kwargs) [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self.wait() [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self[:] = self._gt.wait() [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] return self._exit_event.wait() [ 764.655538] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] current.throw(*self._exc) [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] result = function(*args, **kwargs) [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] return func(*args, **kwargs) [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] raise e [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] nwinfo = self.network_api.allocate_for_instance( [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] created_port_ids = self._update_ports_for_instance( [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.655978] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] with excutils.save_and_reraise_exception(): [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] self.force_reraise() [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] raise self.value [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] updated_port = self._update_port( [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] _ensure_no_port_binding_failure(port) [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] raise exception.PortBindingFailed(port_id=port['id']) [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] nova.exception.PortBindingFailed: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. [ 764.656429] env[62952]: ERROR nova.compute.manager [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] [ 764.656776] env[62952]: DEBUG nova.compute.utils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 764.656776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.944s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.659924] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Build of instance 869c5e7f-b937-4add-ace2-7d9b178c2acc was re-scheduled: Binding failed for port 36f66ae9-b16e-41fe-b029-370e6ec6b611, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 764.660162] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 764.660363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquiring lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.660774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Acquired lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.660774] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.704623] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.729019] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.729281] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.729431] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.729607] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.729752] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.729896] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.730113] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.730285] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.730447] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.730606] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.730772] env[62952]: DEBUG nova.virt.hardware [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.731645] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376e9f93-d70b-4d58-a94e-614390052b66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.739159] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dafa719-7127-4fce-8913-d3dab4c0fc59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.754468] env[62952]: ERROR nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Traceback (most recent call last): [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] yield resources [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self.driver.spawn(context, instance, image_meta, [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] vm_ref = self.build_virtual_machine(instance, [ 764.754468] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] for vif in network_info: [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] return self._sync_wrapper(fn, *args, **kwargs) [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self.wait() [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self[:] = self._gt.wait() [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] return self._exit_event.wait() [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.755240] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] current.throw(*self._exc) [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] result = function(*args, **kwargs) [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] return func(*args, **kwargs) [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] raise e [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] nwinfo = self.network_api.allocate_for_instance( [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] created_port_ids = self._update_ports_for_instance( [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] with excutils.save_and_reraise_exception(): [ 764.756407] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self.force_reraise() [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] raise self.value [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] updated_port = self._update_port( [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] _ensure_no_port_binding_failure(port) [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] raise exception.PortBindingFailed(port_id=port['id']) [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 764.756893] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] [ 764.756893] env[62952]: INFO nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Terminating instance [ 764.757278] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquiring lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.965746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.007198] env[62952]: DEBUG nova.network.neutron [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.140834] env[62952]: DEBUG nova.network.neutron [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.194552] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.323025] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.480480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9207db2b-71fb-4e87-82e1-7a1a8319c4f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.488344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb85c70-ec32-4448-92a8-09cbcdeb5573 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.519054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a245a3ed-6d69-4791-965f-ee1372a1cecc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.526332] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc470ef7-3be7-4eb7-ba4d-b929b898330d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.539483] env[62952]: DEBUG nova.compute.provider_tree [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.645418] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ec57aa-c9fa-4e9a-937e-758f4826f411 req-df0558bb-b9b1-436e-85f2-153d45845f3f service nova] Releasing lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.645839] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquired lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.646040] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.830883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Releasing lock "refresh_cache-869c5e7f-b937-4add-ace2-7d9b178c2acc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.831148] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 765.831397] env[62952]: DEBUG nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.831505] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.967363] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.042980] env[62952]: DEBUG nova.scheduler.client.report [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.177079] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.321212] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.472745] env[62952]: DEBUG nova.network.neutron [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.509504] env[62952]: DEBUG nova.compute.manager [req-233edca5-7dcd-4c08-9cae-da06ea265398 req-af9f4621-343c-476a-950a-4d37565a5136 service nova] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Received event network-vif-deleted-a6e81871-2452-4c69-bfe2-64804255718e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.548872] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.549646] env[62952]: ERROR nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Traceback (most recent call last): [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self.driver.spawn(context, instance, image_meta, [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] vm_ref = self.build_virtual_machine(instance, [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.549646] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] for vif in network_info: [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] return self._sync_wrapper(fn, *args, **kwargs) [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self.wait() [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self[:] = self._gt.wait() [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] return self._exit_event.wait() [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] current.throw(*self._exc) [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.549982] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] result = function(*args, **kwargs) [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] return func(*args, **kwargs) [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] raise e [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] nwinfo = self.network_api.allocate_for_instance( [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] created_port_ids = self._update_ports_for_instance( [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] with excutils.save_and_reraise_exception(): [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] self.force_reraise() [ 766.550459] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] raise self.value [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] updated_port = self._update_port( [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] _ensure_no_port_binding_failure(port) [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] raise exception.PortBindingFailed(port_id=port['id']) [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] nova.exception.PortBindingFailed: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. [ 766.550788] env[62952]: ERROR nova.compute.manager [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] [ 766.550788] env[62952]: DEBUG nova.compute.utils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.553011] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.073s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.555973] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Build of instance ac386ddb-c812-4689-b9b4-b98d26ad1831 was re-scheduled: Binding failed for port 72f3f1b8-4a28-4714-b5f5-ac0bcf514015, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 766.556408] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 766.556631] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquiring lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.556778] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Acquired lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.556949] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.824944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Releasing lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.825401] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.825591] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.825898] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa0bde31-714a-4b8a-8bd6-c35a7e495a84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.835021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6277c7-be83-4228-8b59-bdc7476a8937 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.858221] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc could not be found. [ 766.858444] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.858625] env[62952]: INFO nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Took 0.03 seconds to destroy the instance on the hypervisor. [ 766.858867] env[62952]: DEBUG oslo.service.loopingcall [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.859107] env[62952]: DEBUG nova.compute.manager [-] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.859200] env[62952]: DEBUG nova.network.neutron [-] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.875719] env[62952]: DEBUG nova.network.neutron [-] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.975891] env[62952]: INFO nova.compute.manager [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] [instance: 869c5e7f-b937-4add-ace2-7d9b178c2acc] Took 1.14 seconds to deallocate network for instance. [ 767.094213] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.174484] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.378203] env[62952]: DEBUG nova.network.neutron [-] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.394809] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1016c367-d682-4ddc-970f-8c191015fbf0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.402916] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fd3540-c4e1-4840-baf0-d1dbbf8974bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.433871] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0da0cad-ca4f-4789-a99d-a7ed78853b11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.441135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a738e3f3-353d-4f3d-a925-98935e22d226 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.454668] env[62952]: DEBUG nova.compute.provider_tree [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.681735] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Releasing lock "refresh_cache-ac386ddb-c812-4689-b9b4-b98d26ad1831" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.682078] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 767.682203] env[62952]: DEBUG nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.682377] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.731040] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.882123] env[62952]: INFO nova.compute.manager [-] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Took 1.02 seconds to deallocate network for instance. [ 767.884980] env[62952]: DEBUG nova.compute.claims [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.885183] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.957444] env[62952]: DEBUG nova.scheduler.client.report [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.010710] env[62952]: INFO nova.scheduler.client.report [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Deleted allocations for instance 869c5e7f-b937-4add-ace2-7d9b178c2acc [ 768.233250] env[62952]: DEBUG nova.network.neutron [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.463577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.464280] env[62952]: ERROR nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Traceback (most recent call last): [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self.driver.spawn(context, instance, image_meta, [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] vm_ref = self.build_virtual_machine(instance, [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.464280] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] for vif in network_info: [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] return self._sync_wrapper(fn, *args, **kwargs) [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self.wait() [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self[:] = self._gt.wait() [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] return self._exit_event.wait() [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] current.throw(*self._exc) [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.464680] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] result = function(*args, **kwargs) [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] return func(*args, **kwargs) [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] raise e [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] nwinfo = self.network_api.allocate_for_instance( [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] created_port_ids = self._update_ports_for_instance( [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] with excutils.save_and_reraise_exception(): [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] self.force_reraise() [ 768.465384] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] raise self.value [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] updated_port = self._update_port( [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] _ensure_no_port_binding_failure(port) [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] raise exception.PortBindingFailed(port_id=port['id']) [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] nova.exception.PortBindingFailed: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. [ 768.465784] env[62952]: ERROR nova.compute.manager [instance: f4ec1e94-1778-4296-989d-782700c6faf8] [ 768.465784] env[62952]: DEBUG nova.compute.utils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 768.468503] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.526s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.469934] env[62952]: INFO nova.compute.claims [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.476844] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Build of instance f4ec1e94-1778-4296-989d-782700c6faf8 was re-scheduled: Binding failed for port e00b15fd-a809-45b2-a3ff-92c61ade6abe, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 768.477898] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 768.477898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.477898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.477898] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.520429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0eff9fe4-f860-4fe1-9211-1f949677c509 tempest-ServerActionsV293TestJSON-166165411 tempest-ServerActionsV293TestJSON-166165411-project-member] Lock "869c5e7f-b937-4add-ace2-7d9b178c2acc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.557s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.736223] env[62952]: INFO nova.compute.manager [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] [instance: ac386ddb-c812-4689-b9b4-b98d26ad1831] Took 1.05 seconds to deallocate network for instance. [ 769.009730] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.024537] env[62952]: DEBUG nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.128191] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.555294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.632646] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-f4ec1e94-1778-4296-989d-782700c6faf8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.633898] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 769.633898] env[62952]: DEBUG nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.633898] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.657882] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.775869] env[62952]: INFO nova.scheduler.client.report [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Deleted allocations for instance ac386ddb-c812-4689-b9b4-b98d26ad1831 [ 769.955115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e116c051-75c1-4557-8e62-485df722f7d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.962210] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffad641c-b9ba-460a-9841-fa49eaa3f28b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.995916] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5131d7d4-2c63-4f02-884e-c5a16c2fc07e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.004683] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57bf97f-a67d-429d-9b2a-d26fbf0228c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.017239] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.168709] env[62952]: DEBUG nova.network.neutron [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.282592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquiring lock "69c93772-822a-4f5b-b2d1-95e82a46f286" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.282996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Lock "69c93772-822a-4f5b-b2d1-95e82a46f286" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.290845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fbf663fc-cb6a-4998-85c1-57c27a0d6113 tempest-ListImageFiltersTestJSON-933601881 tempest-ListImageFiltersTestJSON-933601881-project-member] Lock "ac386ddb-c812-4689-b9b4-b98d26ad1831" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.823s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.522021] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.673078] env[62952]: INFO nova.compute.manager [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: f4ec1e94-1778-4296-989d-782700c6faf8] Took 1.04 seconds to deallocate network for instance. [ 770.793272] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.025655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.026345] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.032125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.635s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.326782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.538386] env[62952]: DEBUG nova.compute.utils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.544301] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.544516] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.643229] env[62952]: DEBUG nova.policy [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e4f40cf7ecb417ab616bbc67540006a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7786c2281424cf5a09e49330446f5b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.737486] env[62952]: INFO nova.scheduler.client.report [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleted allocations for instance f4ec1e94-1778-4296-989d-782700c6faf8 [ 771.936707] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242f465f-d83e-4d9d-8358-183be62bdee5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.949532] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959d52a3-615f-4cf2-aeb0-703ff96ec972 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.987202] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72652f8-69e3-4c74-b7b2-ee75bed07889 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.995593] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4381458f-16dc-4449-84b0-2d555c7c62cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.011067] env[62952]: DEBUG nova.compute.provider_tree [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.045131] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.221263] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Successfully created port: 8b5c79ba-cada-40ca-9d4c-a87b3783fb11 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.249746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25aa585b-74d3-43e4-977b-00e9b4ed7b23 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "f4ec1e94-1778-4296-989d-782700c6faf8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.750s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.515206] env[62952]: DEBUG nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.754957] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.028021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.994s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.028021] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Traceback (most recent call last): [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self.driver.spawn(context, instance, image_meta, [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.028021] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] vm_ref = self.build_virtual_machine(instance, [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] for vif in network_info: [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] return self._sync_wrapper(fn, *args, **kwargs) [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self.wait() [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self[:] = self._gt.wait() [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] return self._exit_event.wait() [ 773.028564] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] current.throw(*self._exc) [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] result = function(*args, **kwargs) [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] return func(*args, **kwargs) [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] raise e [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] nwinfo = self.network_api.allocate_for_instance( [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] created_port_ids = self._update_ports_for_instance( [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.028881] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] with excutils.save_and_reraise_exception(): [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] self.force_reraise() [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] raise self.value [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] updated_port = self._update_port( [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] _ensure_no_port_binding_failure(port) [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] raise exception.PortBindingFailed(port_id=port['id']) [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] nova.exception.PortBindingFailed: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. [ 773.029209] env[62952]: ERROR nova.compute.manager [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] [ 773.029504] env[62952]: DEBUG nova.compute.utils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.034084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.876s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.038412] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Build of instance 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71 was re-scheduled: Binding failed for port e4596600-6a36-4ccc-a18f-b514c54f1f33, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 773.040381] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 773.040381] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.040381] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquired lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.040381] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.058388] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.088727] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.089075] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.089298] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.089535] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.089735] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.089950] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.090226] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.090445] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.090667] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.090948] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.091200] env[62952]: DEBUG nova.virt.hardware [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.093149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dffe3ef-1d41-48d3-bb3f-ce5965af30db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.101997] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72970c01-7463-4b75-9fb2-79cbfba64f85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.284398] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.397172] env[62952]: DEBUG nova.compute.manager [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Received event network-changed-8b5c79ba-cada-40ca-9d4c-a87b3783fb11 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.397172] env[62952]: DEBUG nova.compute.manager [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Refreshing instance network info cache due to event network-changed-8b5c79ba-cada-40ca-9d4c-a87b3783fb11. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.397172] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] Acquiring lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.397172] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] Acquired lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.397172] env[62952]: DEBUG nova.network.neutron [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Refreshing network info cache for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 773.521798] env[62952]: ERROR nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 773.521798] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.521798] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.521798] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.521798] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.521798] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.521798] env[62952]: ERROR nova.compute.manager raise self.value [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.521798] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.521798] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.521798] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.522342] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.522342] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.522342] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 773.522342] env[62952]: ERROR nova.compute.manager [ 773.522342] env[62952]: Traceback (most recent call last): [ 773.522342] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.522342] env[62952]: listener.cb(fileno) [ 773.522342] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.522342] env[62952]: result = function(*args, **kwargs) [ 773.522342] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.522342] env[62952]: return func(*args, **kwargs) [ 773.522342] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.522342] env[62952]: raise e [ 773.522342] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.522342] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 773.522342] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.522342] env[62952]: created_port_ids = self._update_ports_for_instance( [ 773.522342] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.522342] env[62952]: with excutils.save_and_reraise_exception(): [ 773.522342] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.522342] env[62952]: self.force_reraise() [ 773.522342] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.522342] env[62952]: raise self.value [ 773.522342] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.522342] env[62952]: updated_port = self._update_port( [ 773.522342] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.522342] env[62952]: _ensure_no_port_binding_failure(port) [ 773.522342] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.522342] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.523282] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 773.523282] env[62952]: Removing descriptor: 15 [ 773.523282] env[62952]: ERROR nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Traceback (most recent call last): [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] yield resources [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self.driver.spawn(context, instance, image_meta, [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.523282] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] vm_ref = self.build_virtual_machine(instance, [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] for vif in network_info: [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return self._sync_wrapper(fn, *args, **kwargs) [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self.wait() [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self[:] = self._gt.wait() [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return self._exit_event.wait() [ 773.523665] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] result = hub.switch() [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return self.greenlet.switch() [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] result = function(*args, **kwargs) [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return func(*args, **kwargs) [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] raise e [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] nwinfo = self.network_api.allocate_for_instance( [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.524103] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] created_port_ids = self._update_ports_for_instance( [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] with excutils.save_and_reraise_exception(): [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self.force_reraise() [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] raise self.value [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] updated_port = self._update_port( [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] _ensure_no_port_binding_failure(port) [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.524501] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] raise exception.PortBindingFailed(port_id=port['id']) [ 773.524863] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 773.524863] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] [ 773.524863] env[62952]: INFO nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Terminating instance [ 773.528182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquiring lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.567295] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.715190] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.922944] env[62952]: DEBUG nova.network.neutron [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.933241] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6abc9c6-8b6b-4dfb-a80d-375caa04129a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.941699] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fad592-1aa0-464e-b2aa-a23b2282faf4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.975381] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b00fbaa-2a12-4f78-a699-951531a42b42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.984449] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3fca97-6596-4612-bee6-294cb7877c4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.998692] env[62952]: DEBUG nova.compute.provider_tree [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.220041] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Releasing lock "refresh_cache-4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.220041] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 774.220041] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.220041] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.250061] env[62952]: DEBUG nova.network.neutron [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.257178] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.505255] env[62952]: DEBUG nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.758187] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b72ea05-ce57-438b-81ed-04a266435405 req-7627e021-ae18-40f5-9593-aa4bc1d7b77e service nova] Releasing lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.758792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquired lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.758792] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.763887] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.010517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.011172] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Traceback (most recent call last): [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self.driver.spawn(context, instance, image_meta, [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] vm_ref = self.build_virtual_machine(instance, [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.011172] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] for vif in network_info: [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return self._sync_wrapper(fn, *args, **kwargs) [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self.wait() [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self[:] = self._gt.wait() [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return self._exit_event.wait() [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] result = hub.switch() [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 775.011506] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return self.greenlet.switch() [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] result = function(*args, **kwargs) [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] return func(*args, **kwargs) [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] raise e [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] nwinfo = self.network_api.allocate_for_instance( [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] created_port_ids = self._update_ports_for_instance( [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] with excutils.save_and_reraise_exception(): [ 775.011847] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] self.force_reraise() [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] raise self.value [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] updated_port = self._update_port( [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] _ensure_no_port_binding_failure(port) [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] raise exception.PortBindingFailed(port_id=port['id']) [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] nova.exception.PortBindingFailed: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. [ 775.012221] env[62952]: ERROR nova.compute.manager [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] [ 775.012581] env[62952]: DEBUG nova.compute.utils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.013699] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.336s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.022177] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Build of instance a825d324-0b26-4051-bdfa-4a4246f6d4e9 was re-scheduled: Binding failed for port b1f7df5e-c0d9-494e-8185-b8f132bf3a44, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.022672] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.022903] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.025679] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquired lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.025936] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.269112] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71] Took 1.05 seconds to deallocate network for instance. [ 775.279687] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.424109] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.547057] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.636958] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.807154] env[62952]: DEBUG nova.compute.manager [req-2baccea6-988f-4d73-9476-00656156be5d req-010e3411-4551-4fed-beca-edad0ef29cb9 service nova] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Received event network-vif-deleted-8b5c79ba-cada-40ca-9d4c-a87b3783fb11 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.868416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "62d47ebb-8782-4062-a27e-556c758ed475" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.869087] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "62d47ebb-8782-4062-a27e-556c758ed475" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.893679] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.894612] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.894612] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 775.894792] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 775.927045] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Releasing lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.927463] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.927664] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.929057] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bad60999-b017-447d-b651-f13b41a3234c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.937783] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0b7b79-548b-4dae-9b1a-aee12f0f00bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.967856] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 38e05fe2-d938-43d6-ab2b-b99fb7be51d5 could not be found. [ 775.968276] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.968676] env[62952]: INFO nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 775.969031] env[62952]: DEBUG oslo.service.loopingcall [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.972034] env[62952]: DEBUG nova.compute.manager [-] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.972133] env[62952]: DEBUG nova.network.neutron [-] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.988757] env[62952]: DEBUG nova.network.neutron [-] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.014207] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39240e4d-51bb-4f30-8b69-1390fa9d7c8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.022999] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c7b7a1-aad2-46d9-8a47-d06f7b482218 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.056703] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3460aad7-a12c-4d7d-9872-0ac8233f20a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.064772] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2be5ee-f001-4112-a5fc-6357c1d2c7b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.082042] env[62952]: DEBUG nova.compute.provider_tree [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.139689] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Releasing lock "refresh_cache-a825d324-0b26-4051-bdfa-4a4246f6d4e9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.139935] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.140131] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.140303] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.160013] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquiring lock "56aa8186-2bcb-47fc-a832-31654cbf218b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.160310] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Lock "56aa8186-2bcb-47fc-a832-31654cbf218b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.161197] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.243985] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "351b1672-ddb3-4562-9965-d65a7d869e82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.244259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.300898] env[62952]: INFO nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Deleted allocations for instance 4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71 [ 776.398402] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 776.398624] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 776.398957] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 776.398957] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.399128] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.399255] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.399394] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.399540] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.399684] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.399807] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 776.399948] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.489893] env[62952]: DEBUG nova.network.neutron [-] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.586269] env[62952]: DEBUG nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.663767] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.808737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "4ea5ebc8-c0d5-43c0-a864-d8cf7efdee71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.487s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.902883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.992026] env[62952]: INFO nova.compute.manager [-] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Took 1.02 seconds to deallocate network for instance. [ 776.995320] env[62952]: DEBUG nova.compute.claims [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.995587] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.091577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.078s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.093022] env[62952]: ERROR nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Traceback (most recent call last): [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self.driver.spawn(context, instance, image_meta, [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] vm_ref = self.build_virtual_machine(instance, [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.093022] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] for vif in network_info: [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] return self._sync_wrapper(fn, *args, **kwargs) [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self.wait() [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self[:] = self._gt.wait() [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] return self._exit_event.wait() [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] current.throw(*self._exc) [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.093448] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] result = function(*args, **kwargs) [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] return func(*args, **kwargs) [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] raise e [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] nwinfo = self.network_api.allocate_for_instance( [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] created_port_ids = self._update_ports_for_instance( [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] with excutils.save_and_reraise_exception(): [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] self.force_reraise() [ 777.093858] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] raise self.value [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] updated_port = self._update_port( [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] _ensure_no_port_binding_failure(port) [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] raise exception.PortBindingFailed(port_id=port['id']) [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] nova.exception.PortBindingFailed: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. [ 777.094244] env[62952]: ERROR nova.compute.manager [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] [ 777.094244] env[62952]: DEBUG nova.compute.utils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.097335] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.032s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.098377] env[62952]: INFO nova.compute.claims [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.101037] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Build of instance 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec was re-scheduled: Binding failed for port 1463b38b-65ad-49e0-a75d-3dd3e86e3438, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.101464] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.101688] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquiring lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.101830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Acquired lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.101981] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.166635] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: a825d324-0b26-4051-bdfa-4a4246f6d4e9] Took 1.03 seconds to deallocate network for instance. [ 777.313321] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.623368] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.708178] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.840311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.194587] env[62952]: INFO nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Deleted allocations for instance a825d324-0b26-4051-bdfa-4a4246f6d4e9 [ 778.210729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Releasing lock "refresh_cache-04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.210991] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.211218] env[62952]: DEBUG nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.214021] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.230484] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.448017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72a352e-0c62-49b3-9f7c-043bc41a8b10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.456526] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a268468a-9ad5-4916-b74d-c3ff13268f0f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.484843] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffeae6d-28bf-4dc3-9cf7-faef5fa2d502 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.491325] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5658e1-c138-406f-a32f-b67c462faa03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.503995] env[62952]: DEBUG nova.compute.provider_tree [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.705178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "a825d324-0b26-4051-bdfa-4a4246f6d4e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.347s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.734925] env[62952]: DEBUG nova.network.neutron [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.009054] env[62952]: DEBUG nova.scheduler.client.report [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.208142] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.237795] env[62952]: INFO nova.compute.manager [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] [instance: 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec] Took 1.03 seconds to deallocate network for instance. [ 779.512980] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.513534] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.516241] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.551s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.517641] env[62952]: INFO nova.compute.claims [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.730589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.023536] env[62952]: DEBUG nova.compute.utils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.025604] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.025894] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 780.086912] env[62952]: DEBUG nova.policy [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4e5d87441d1430f84b10a8397b4834e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9908ca9c44c41c7a0cfe9e1af34a4aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.265660] env[62952]: INFO nova.scheduler.client.report [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Deleted allocations for instance 04f5dd28-9b5e-4175-9f00-bbd55e8db2ec [ 780.439290] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Successfully created port: 937f84a5-94cd-42da-a5de-53578453551b {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.526799] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.720360] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Successfully created port: 3fe9f99d-c631-4608-9325-77faa55db814 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.774096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a4b9929d-3b0f-42cf-9100-f0db9e66d073 tempest-ListServersNegativeTestJSON-908730112 tempest-ListServersNegativeTestJSON-908730112-project-member] Lock "04f5dd28-9b5e-4175-9f00-bbd55e8db2ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.391s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.835893] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89540e49-b6cb-4cc4-9926-ab14d51644af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.843655] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6999ab26-4b54-47a5-a87b-4b958da90e43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.875972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da01099f-f5b4-43bc-888c-6d6f385867b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.883937] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fed0a8-ef83-4e62-81e9-c8ba42b89fca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.897890] env[62952]: DEBUG nova.compute.provider_tree [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.276969] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.401083] env[62952]: DEBUG nova.scheduler.client.report [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.538274] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.565711] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.565957] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.566127] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.566311] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.566455] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.567902] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.568223] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.568440] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.568658] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.568922] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.569179] env[62952]: DEBUG nova.virt.hardware [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.570091] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a794878-c1f2-4841-baf3-0a2921189ce3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.578494] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed384f63-b895-4ecc-a64a-8c9e628d99e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.635671] env[62952]: DEBUG nova.compute.manager [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Received event network-changed-937f84a5-94cd-42da-a5de-53578453551b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.635889] env[62952]: DEBUG nova.compute.manager [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Refreshing instance network info cache due to event network-changed-937f84a5-94cd-42da-a5de-53578453551b. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.636113] env[62952]: DEBUG oslo_concurrency.lockutils [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] Acquiring lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.636258] env[62952]: DEBUG oslo_concurrency.lockutils [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] Acquired lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.636417] env[62952]: DEBUG nova.network.neutron [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Refreshing network info cache for port 937f84a5-94cd-42da-a5de-53578453551b {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.716536] env[62952]: ERROR nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 781.716536] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.716536] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.716536] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.716536] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.716536] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.716536] env[62952]: ERROR nova.compute.manager raise self.value [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.716536] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.716536] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.716536] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.717091] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.717091] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.717091] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 781.717091] env[62952]: ERROR nova.compute.manager [ 781.717091] env[62952]: Traceback (most recent call last): [ 781.717091] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.717091] env[62952]: listener.cb(fileno) [ 781.717091] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.717091] env[62952]: result = function(*args, **kwargs) [ 781.717091] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.717091] env[62952]: return func(*args, **kwargs) [ 781.717091] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.717091] env[62952]: raise e [ 781.717091] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.717091] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 781.717091] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.717091] env[62952]: created_port_ids = self._update_ports_for_instance( [ 781.717091] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.717091] env[62952]: with excutils.save_and_reraise_exception(): [ 781.717091] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.717091] env[62952]: self.force_reraise() [ 781.717091] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.717091] env[62952]: raise self.value [ 781.717091] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.717091] env[62952]: updated_port = self._update_port( [ 781.717091] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.717091] env[62952]: _ensure_no_port_binding_failure(port) [ 781.717091] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.717091] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.717929] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 781.717929] env[62952]: Removing descriptor: 19 [ 781.717929] env[62952]: ERROR nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Traceback (most recent call last): [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] yield resources [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self.driver.spawn(context, instance, image_meta, [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.717929] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] vm_ref = self.build_virtual_machine(instance, [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] for vif in network_info: [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return self._sync_wrapper(fn, *args, **kwargs) [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self.wait() [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self[:] = self._gt.wait() [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return self._exit_event.wait() [ 781.718312] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] result = hub.switch() [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return self.greenlet.switch() [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] result = function(*args, **kwargs) [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return func(*args, **kwargs) [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] raise e [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] nwinfo = self.network_api.allocate_for_instance( [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.718712] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] created_port_ids = self._update_ports_for_instance( [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] with excutils.save_and_reraise_exception(): [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self.force_reraise() [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] raise self.value [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] updated_port = self._update_port( [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] _ensure_no_port_binding_failure(port) [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.719110] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] raise exception.PortBindingFailed(port_id=port['id']) [ 781.719460] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 781.719460] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] [ 781.719460] env[62952]: INFO nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Terminating instance [ 781.719955] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.807289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.909012] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.909564] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 781.911989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.027s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.184377] env[62952]: DEBUG nova.network.neutron [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.329378] env[62952]: DEBUG nova.network.neutron [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.417160] env[62952]: DEBUG nova.compute.utils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.418605] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 782.418834] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.482909] env[62952]: DEBUG nova.policy [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '760c43a867c74a13a54be25d82769e37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea3de73e23ba4a61970b4bffb9e7b5f3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.781258] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f725c9-decb-421b-a132-9cd67cd2870d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.791024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8e5e94-3e5b-409c-933f-1b39813cd111 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.822900] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21631e0-5e94-4d08-8abd-5c14d7fc66f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.830818] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1d0eb9-a512-4ec3-8801-893cb2b349af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.835318] env[62952]: DEBUG oslo_concurrency.lockutils [req-c1034d08-e3d4-4e4f-a527-281f07c39cec req-0f1bd2f5-00cd-4bc3-8472-7d62429a6514 service nova] Releasing lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.835711] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquired lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.835889] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.847481] env[62952]: DEBUG nova.compute.provider_tree [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.928591] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.195973] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Successfully created port: 272b9bc4-bacd-48d8-b805-6a6ce692bafe {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.350463] env[62952]: DEBUG nova.scheduler.client.report [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.414546] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.634411] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.692275] env[62952]: DEBUG nova.compute.manager [req-3e6a1802-248c-4a26-86a9-2d8e88ecef83 req-9dc93ced-772c-4f36-b350-71e450634b10 service nova] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Received event network-vif-deleted-937f84a5-94cd-42da-a5de-53578453551b {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.859315] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.859315] env[62952]: ERROR nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Traceback (most recent call last): [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self.driver.spawn(context, instance, image_meta, [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.859315] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] vm_ref = self.build_virtual_machine(instance, [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] for vif in network_info: [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] return self._sync_wrapper(fn, *args, **kwargs) [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self.wait() [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self[:] = self._gt.wait() [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] return self._exit_event.wait() [ 783.859837] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] current.throw(*self._exc) [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] result = function(*args, **kwargs) [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] return func(*args, **kwargs) [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] raise e [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] nwinfo = self.network_api.allocate_for_instance( [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] created_port_ids = self._update_ports_for_instance( [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.860232] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] with excutils.save_and_reraise_exception(): [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] self.force_reraise() [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] raise self.value [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] updated_port = self._update_port( [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] _ensure_no_port_binding_failure(port) [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] raise exception.PortBindingFailed(port_id=port['id']) [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] nova.exception.PortBindingFailed: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. [ 783.860612] env[62952]: ERROR nova.compute.manager [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] [ 783.860967] env[62952]: DEBUG nova.compute.utils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.864373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.307s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.864373] env[62952]: INFO nova.compute.claims [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.867577] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Build of instance 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc was re-scheduled: Binding failed for port a6e81871-2452-4c69-bfe2-64804255718e, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.868026] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.868613] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquiring lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.868613] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Acquired lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.868613] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.938436] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 783.979882] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.981072] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.981072] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.981072] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.981072] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.981072] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.981353] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.981353] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.981353] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.981450] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.981560] env[62952]: DEBUG nova.virt.hardware [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.982447] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3e4855-53c7-430d-9ced-1f3e82ab3977 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.990905] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd7a685-ffe9-4936-8f40-f70474e7fab2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.137327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Releasing lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.137763] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.137962] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.138278] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a84a356-61d4-4425-93a2-a022698279a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.153451] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beeb764c-42bf-4bcf-a9fd-eee2635b90e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.179983] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5283bf79-bf20-4a38-912f-ced401fbd691 could not be found. [ 784.179983] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.179983] env[62952]: INFO nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Took 0.04 seconds to destroy the instance on the hypervisor. [ 784.179983] env[62952]: DEBUG oslo.service.loopingcall [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.180303] env[62952]: DEBUG nova.compute.manager [-] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.180342] env[62952]: DEBUG nova.network.neutron [-] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.222371] env[62952]: DEBUG nova.network.neutron [-] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.402357] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.536433] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.837688] env[62952]: ERROR nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 784.837688] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.837688] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.837688] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.837688] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.837688] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.837688] env[62952]: ERROR nova.compute.manager raise self.value [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.837688] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 784.837688] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.837688] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 784.838393] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.838393] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 784.838393] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 784.838393] env[62952]: ERROR nova.compute.manager [ 784.838393] env[62952]: Traceback (most recent call last): [ 784.838393] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 784.838393] env[62952]: listener.cb(fileno) [ 784.838393] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.838393] env[62952]: result = function(*args, **kwargs) [ 784.838393] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 784.838393] env[62952]: return func(*args, **kwargs) [ 784.838393] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.838393] env[62952]: raise e [ 784.838393] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.838393] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 784.838393] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.838393] env[62952]: created_port_ids = self._update_ports_for_instance( [ 784.838393] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.838393] env[62952]: with excutils.save_and_reraise_exception(): [ 784.838393] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.838393] env[62952]: self.force_reraise() [ 784.838393] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.838393] env[62952]: raise self.value [ 784.838393] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.838393] env[62952]: updated_port = self._update_port( [ 784.838393] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.838393] env[62952]: _ensure_no_port_binding_failure(port) [ 784.838393] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.838393] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 784.839255] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 784.839255] env[62952]: Removing descriptor: 19 [ 784.839255] env[62952]: ERROR nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Traceback (most recent call last): [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] yield resources [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self.driver.spawn(context, instance, image_meta, [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self._vmops.spawn(context, instance, image_meta, injected_files, [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 784.839255] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] vm_ref = self.build_virtual_machine(instance, [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] vif_infos = vmwarevif.get_vif_info(self._session, [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] for vif in network_info: [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return self._sync_wrapper(fn, *args, **kwargs) [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self.wait() [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self[:] = self._gt.wait() [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return self._exit_event.wait() [ 784.839588] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] result = hub.switch() [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return self.greenlet.switch() [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] result = function(*args, **kwargs) [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return func(*args, **kwargs) [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] raise e [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] nwinfo = self.network_api.allocate_for_instance( [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.839941] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] created_port_ids = self._update_ports_for_instance( [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] with excutils.save_and_reraise_exception(): [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self.force_reraise() [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] raise self.value [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] updated_port = self._update_port( [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] _ensure_no_port_binding_failure(port) [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.840654] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] raise exception.PortBindingFailed(port_id=port['id']) [ 784.841908] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 784.841908] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] [ 784.841908] env[62952]: INFO nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Terminating instance [ 784.841908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquiring lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.841908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquired lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.841908] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.039116] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Releasing lock "refresh_cache-9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.039116] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 785.039116] env[62952]: DEBUG nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.039116] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.058712] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.059996] env[62952]: DEBUG nova.network.neutron [-] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.158974] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881d6eeb-72ec-47aa-a5d2-7cdc37123a56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.166512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b80719-08d1-4424-9029-9f9712eebb43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.198097] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40123f85-a939-4d69-b2bd-05c02c3a34ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.206636] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4fa9a4-9de5-4248-92ad-3629ae390fb7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.220510] env[62952]: DEBUG nova.compute.provider_tree [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.370052] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.472704] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.563128] env[62952]: DEBUG nova.network.neutron [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.564021] env[62952]: INFO nova.compute.manager [-] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Took 1.38 seconds to deallocate network for instance. [ 785.566738] env[62952]: DEBUG nova.compute.claims [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 785.566738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.722884] env[62952]: DEBUG nova.scheduler.client.report [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.757297] env[62952]: DEBUG nova.compute.manager [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Received event network-changed-272b9bc4-bacd-48d8-b805-6a6ce692bafe {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.757525] env[62952]: DEBUG nova.compute.manager [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Refreshing instance network info cache due to event network-changed-272b9bc4-bacd-48d8-b805-6a6ce692bafe. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.757726] env[62952]: DEBUG oslo_concurrency.lockutils [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] Acquiring lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.976656] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Releasing lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.977137] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.977295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.978086] env[62952]: DEBUG oslo_concurrency.lockutils [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] Acquired lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.978086] env[62952]: DEBUG nova.network.neutron [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Refreshing network info cache for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.981429] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c491baf-3c51-4bbf-ae64-d00e8f8d1116 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.987738] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1463567f-0a5f-4302-b855-2b8e2675653f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.008596] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fc5b891b-8432-4b2f-97d9-c93ca0a18783 could not be found. [ 786.008794] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.008974] env[62952]: INFO nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Took 0.03 seconds to destroy the instance on the hypervisor. [ 786.009214] env[62952]: DEBUG oslo.service.loopingcall [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.009412] env[62952]: DEBUG nova.compute.manager [-] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.009504] env[62952]: DEBUG nova.network.neutron [-] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.023339] env[62952]: DEBUG nova.network.neutron [-] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.066399] env[62952]: INFO nova.compute.manager [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] [instance: 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc] Took 1.03 seconds to deallocate network for instance. [ 786.228311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.228899] env[62952]: DEBUG nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.232647] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.905s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.233886] env[62952]: INFO nova.compute.claims [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.503174] env[62952]: DEBUG nova.network.neutron [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.525411] env[62952]: DEBUG nova.network.neutron [-] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.601950] env[62952]: DEBUG nova.network.neutron [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.737830] env[62952]: DEBUG nova.compute.utils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.743385] env[62952]: DEBUG nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 787.028106] env[62952]: INFO nova.compute.manager [-] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Took 1.02 seconds to deallocate network for instance. [ 787.030238] env[62952]: DEBUG nova.compute.claims [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 787.030409] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.103798] env[62952]: DEBUG oslo_concurrency.lockutils [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] Releasing lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.104098] env[62952]: DEBUG nova.compute.manager [req-1295e8e4-b834-4d4b-b360-8738830a6f9b req-611dca26-233b-4b4c-8bc4-253388550a39 service nova] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Received event network-vif-deleted-272b9bc4-bacd-48d8-b805-6a6ce692bafe {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.116399] env[62952]: INFO nova.scheduler.client.report [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Deleted allocations for instance 9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc [ 787.241847] env[62952]: DEBUG nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.595165] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad482678-e7dc-4c10-886b-52b77bb21bd8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.603129] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c995fd-cc23-49e2-b3df-ce33ab27c384 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.633124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5cc93b04-8a11-4177-98c8-e522b6e72aa0 tempest-ServersTestJSON-1581970525 tempest-ServersTestJSON-1581970525-project-member] Lock "9f6e1fdf-d7b3-48eb-b4f9-729d4c1fb9cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.406s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.635036] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66cea612-327a-48ea-ba5a-940dd85951dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.642994] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93dd2bca-ce34-4489-a7d4-19b36bf64a0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.659723] env[62952]: DEBUG nova.compute.provider_tree [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.007423] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.007542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.138907] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.164458] env[62952]: DEBUG nova.scheduler.client.report [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.253447] env[62952]: DEBUG nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.280366] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.280609] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.280764] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.280942] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.281103] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.281297] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.281513] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.281669] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.281832] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.281992] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.282229] env[62952]: DEBUG nova.virt.hardware [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.283123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16790e9c-aeb7-4dc1-bffc-634c035ee3f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.292516] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fba1253-0869-47c1-9a1f-cad399f72692 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.304758] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.310364] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Creating folder: Project (bcb5d4bdf83a4d159128e030c6d12ae8). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.310739] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e33a5ca6-ee98-4135-a460-fb446503af9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.320954] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Created folder: Project (bcb5d4bdf83a4d159128e030c6d12ae8) in parent group-v290852. [ 788.320954] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Creating folder: Instances. Parent ref: group-v290866. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.320954] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-51939c4c-8922-4c0d-82ad-4b4880d2a4ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.329308] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Created folder: Instances in parent group-v290866. [ 788.329308] env[62952]: DEBUG oslo.service.loopingcall [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.329517] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.329701] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b213cec-8a8b-42ad-a507-7322667edee8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.346251] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.346251] env[62952]: value = "task-1367073" [ 788.346251] env[62952]: _type = "Task" [ 788.346251] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.354885] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367073, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.662813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.669814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.670335] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.672904] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.390s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.676120] env[62952]: INFO nova.compute.claims [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.856411] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367073, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.179092] env[62952]: DEBUG nova.compute.utils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.182197] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.182369] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.236958] env[62952]: DEBUG nova.policy [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '301bff8b7e424da8a71aebf8d3137568', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '820b057a4e71403eb0d244011cfc4e55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.356751] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367073, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.685549] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.704256] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Successfully created port: 3af034e9-59a9-4803-8390-ec339b7ae0d5 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.858796] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367073, 'name': CreateVM_Task, 'duration_secs': 1.257844} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.861474] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.862089] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.862343] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.862695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.862952] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcb518b6-517c-415a-ab43-8e2354f3d867 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.867665] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 789.867665] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c1df38-16c1-10c0-d0e0-f0dbe3f6ccc1" [ 789.867665] env[62952]: _type = "Task" [ 789.867665] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.877621] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c1df38-16c1-10c0-d0e0-f0dbe3f6ccc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.087870] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592f5b49-cd5e-4716-a3b1-b21ff9e64f9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.097179] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f975f5-76ad-43cd-a8d3-aee3cf138faa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.129295] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4914950e-d221-41eb-b082-30c8d5040896 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.136810] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a634946b-300b-469a-b772-dcafea6f1a7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.158277] env[62952]: DEBUG nova.compute.provider_tree [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.377758] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c1df38-16c1-10c0-d0e0-f0dbe3f6ccc1, 'name': SearchDatastore_Task, 'duration_secs': 0.008982} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.378112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.378441] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.378822] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.379072] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.379636] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.379882] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f3a687f-d9ab-4c25-8489-9410aa4db0a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.388223] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.388408] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.389116] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17e495f6-29c1-4a53-a481-a195af25d249 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.394238] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 790.394238] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526faddd-2d29-8bfa-20c1-4b5254e166b1" [ 790.394238] env[62952]: _type = "Task" [ 790.394238] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.401716] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526faddd-2d29-8bfa-20c1-4b5254e166b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.664947] env[62952]: DEBUG nova.scheduler.client.report [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.701382] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.731240] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.737734] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.737734] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.737734] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.737734] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.737734] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.738080] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.738080] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.738080] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.738080] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.738254] env[62952]: DEBUG nova.virt.hardware [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.740275] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c26182e-d015-4e2c-9d0b-6221f86669af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.748857] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406b6fce-5d19-4568-8c95-e3fffeb35b84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.904761] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526faddd-2d29-8bfa-20c1-4b5254e166b1, 'name': SearchDatastore_Task, 'duration_secs': 0.007683} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.905732] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-399ba224-7c59-41ce-aa59-29fee618097a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.910846] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 790.910846] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]525cdb04-b9c7-0697-2f03-069390c9e959" [ 790.910846] env[62952]: _type = "Task" [ 790.910846] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.919123] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525cdb04-b9c7-0697-2f03-069390c9e959, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.953260] env[62952]: DEBUG nova.compute.manager [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Received event network-changed-3af034e9-59a9-4803-8390-ec339b7ae0d5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.953260] env[62952]: DEBUG nova.compute.manager [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Refreshing instance network info cache due to event network-changed-3af034e9-59a9-4803-8390-ec339b7ae0d5. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.953260] env[62952]: DEBUG oslo_concurrency.lockutils [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] Acquiring lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.953260] env[62952]: DEBUG oslo_concurrency.lockutils [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] Acquired lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.953260] env[62952]: DEBUG nova.network.neutron [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Refreshing network info cache for port 3af034e9-59a9-4803-8390-ec339b7ae0d5 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.148216] env[62952]: ERROR nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 791.148216] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.148216] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.148216] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.148216] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.148216] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.148216] env[62952]: ERROR nova.compute.manager raise self.value [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.148216] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 791.148216] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.148216] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 791.148668] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.148668] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 791.148668] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 791.148668] env[62952]: ERROR nova.compute.manager [ 791.148668] env[62952]: Traceback (most recent call last): [ 791.148668] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 791.148668] env[62952]: listener.cb(fileno) [ 791.148668] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.148668] env[62952]: result = function(*args, **kwargs) [ 791.148668] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.148668] env[62952]: return func(*args, **kwargs) [ 791.148668] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.148668] env[62952]: raise e [ 791.148668] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.148668] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 791.148668] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.148668] env[62952]: created_port_ids = self._update_ports_for_instance( [ 791.148668] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.148668] env[62952]: with excutils.save_and_reraise_exception(): [ 791.148668] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.148668] env[62952]: self.force_reraise() [ 791.148668] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.148668] env[62952]: raise self.value [ 791.148668] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.148668] env[62952]: updated_port = self._update_port( [ 791.148668] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.148668] env[62952]: _ensure_no_port_binding_failure(port) [ 791.148668] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.148668] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 791.149528] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 791.149528] env[62952]: Removing descriptor: 19 [ 791.149528] env[62952]: ERROR nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Traceback (most recent call last): [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] yield resources [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self.driver.spawn(context, instance, image_meta, [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.149528] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] vm_ref = self.build_virtual_machine(instance, [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] for vif in network_info: [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return self._sync_wrapper(fn, *args, **kwargs) [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self.wait() [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self[:] = self._gt.wait() [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return self._exit_event.wait() [ 791.149857] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] result = hub.switch() [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return self.greenlet.switch() [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] result = function(*args, **kwargs) [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return func(*args, **kwargs) [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] raise e [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] nwinfo = self.network_api.allocate_for_instance( [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.150609] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] created_port_ids = self._update_ports_for_instance( [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] with excutils.save_and_reraise_exception(): [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self.force_reraise() [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] raise self.value [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] updated_port = self._update_port( [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] _ensure_no_port_binding_failure(port) [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.151169] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] raise exception.PortBindingFailed(port_id=port['id']) [ 791.151598] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 791.151598] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] [ 791.151598] env[62952]: INFO nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Terminating instance [ 791.151806] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.170916] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.171530] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.173998] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.271s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.174228] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.174370] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 791.174650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.179s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.178134] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5d88a9-83b0-4a1d-b74e-ecb5d7fb03a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.189253] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f8ef30-2d5c-49ba-80f2-809f460752ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.206645] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a2cd7d-118d-4dce-9180-05b24840f99d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.212418] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ccceab-577e-4287-84ac-73da9ce2690f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.249487] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181488MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 791.249909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.421161] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525cdb04-b9c7-0697-2f03-069390c9e959, 'name': SearchDatastore_Task, 'duration_secs': 0.008861} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.421429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.421683] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.421934] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b2db8fc-a414-4e3b-aa76-e2fa782b9c64 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.429596] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 791.429596] env[62952]: value = "task-1367074" [ 791.429596] env[62952]: _type = "Task" [ 791.429596] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.437367] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367074, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.472389] env[62952]: DEBUG nova.network.neutron [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.548550] env[62952]: DEBUG nova.network.neutron [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.679522] env[62952]: DEBUG nova.compute.utils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.681596] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.681863] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.760038] env[62952]: DEBUG nova.policy [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '301bff8b7e424da8a71aebf8d3137568', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '820b057a4e71403eb0d244011cfc4e55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.945869] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367074, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.051352] env[62952]: DEBUG oslo_concurrency.lockutils [req-d19113a9-6b7a-41b3-a71c-c302c217111b req-654185d0-695e-4d02-be33-e9b15b730f55 service nova] Releasing lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.055367] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquired lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.055566] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.089321] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7b57e2-dd05-4b0e-ac10-3bfd440f01f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.097662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5293e85d-74df-47f1-a4e6-dc66de1cb242 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.131033] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Successfully created port: 5a2f6f40-4ad4-4214-b7de-c4538335eaa6 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.133421] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb815d6e-989a-4f50-bc1e-1eb11e91b777 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.146029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e79959e-acc0-4d95-9a6a-45328d6af132 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.162946] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.187022] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.441767] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367074, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549546} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.442045] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.442285] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.442534] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb28d117-61c5-420a-bb1e-476db3cd27c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.448841] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 792.448841] env[62952]: value = "task-1367075" [ 792.448841] env[62952]: _type = "Task" [ 792.448841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.456955] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367075, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.583157] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.702465] env[62952]: ERROR nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [req-2aec3030-a928-4e36-924d-31fa87da9904] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2aec3030-a928-4e36-924d-31fa87da9904"}]}: nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 792.719757] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.723632] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 792.745648] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 792.745648] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.756763] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 792.785350] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 792.841796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "a63ec2ac-4484-4360-962b-105a5338c1b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.842107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.968050] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367075, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067405} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.968050] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.969273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981d7c16-95bf-4ffa-85c3-f35c1bb45c7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.992104] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.995708] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb2778ae-ede0-4091-81fc-be09e7f1c7e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.021358] env[62952]: DEBUG nova.compute.manager [req-c2dd3ba0-9bbe-4842-b173-cde14b0c6172 req-93729603-b2a3-43a7-96f8-3d9f8a2d5e0d service nova] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Received event network-vif-deleted-3af034e9-59a9-4803-8390-ec339b7ae0d5 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.024888] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 793.024888] env[62952]: value = "task-1367076" [ 793.024888] env[62952]: _type = "Task" [ 793.024888] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.036852] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367076, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.059454] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "802fd770-38aa-4417-b25e-fd83510bceef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.059713] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "802fd770-38aa-4417-b25e-fd83510bceef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.200126] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.215796] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc7839c-4931-42db-aa84-f0d2c370e0a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.223239] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Releasing lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.223895] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.223895] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.224738] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d72b7b-8840-41fb-8de0-3f99d4a33fea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.230524] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35a83bd8-1fb6-46d9-9f1f-ce30353db8b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.267092] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.267092] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.267344] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.267344] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.267436] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.267580] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.267784] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.267937] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.268114] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.268274] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.268445] env[62952]: DEBUG nova.virt.hardware [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.269299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95372e99-9ec6-46e3-bb76-6578a552d0f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.276098] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4127316e-dfb1-480d-896f-77994623c33d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.277832] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3a51f3-792e-49c3-8760-481ad7e807bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.303034] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0194f0b6-7154-48c3-82a7-55dc8bb47a1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.310647] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffa0b57-6a4a-4bc8-ae58-adc6ad88fb41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.317876] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f8162d7-c5e6-4c66-8a10-bb06998738b5 could not be found. [ 793.318164] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.318376] env[62952]: INFO nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Took 0.09 seconds to destroy the instance on the hypervisor. [ 793.318692] env[62952]: DEBUG oslo.service.loopingcall [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.334687] env[62952]: DEBUG nova.compute.manager [-] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.334800] env[62952]: DEBUG nova.network.neutron [-] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.336992] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 793.339717] env[62952]: ERROR nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 793.339717] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.339717] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.339717] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.339717] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.339717] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.339717] env[62952]: ERROR nova.compute.manager raise self.value [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.339717] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 793.339717] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.339717] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 793.340275] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.340275] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 793.340275] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 793.340275] env[62952]: ERROR nova.compute.manager [ 793.340275] env[62952]: Traceback (most recent call last): [ 793.340275] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 793.340275] env[62952]: listener.cb(fileno) [ 793.340275] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.340275] env[62952]: result = function(*args, **kwargs) [ 793.340275] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.340275] env[62952]: return func(*args, **kwargs) [ 793.340275] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.340275] env[62952]: raise e [ 793.340275] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.340275] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 793.340275] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.340275] env[62952]: created_port_ids = self._update_ports_for_instance( [ 793.340275] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.340275] env[62952]: with excutils.save_and_reraise_exception(): [ 793.340275] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.340275] env[62952]: self.force_reraise() [ 793.340275] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.340275] env[62952]: raise self.value [ 793.340275] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.340275] env[62952]: updated_port = self._update_port( [ 793.340275] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.340275] env[62952]: _ensure_no_port_binding_failure(port) [ 793.340275] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.340275] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 793.341237] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 793.341237] env[62952]: Removing descriptor: 19 [ 793.341237] env[62952]: ERROR nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Traceback (most recent call last): [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] yield resources [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self.driver.spawn(context, instance, image_meta, [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.341237] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] vm_ref = self.build_virtual_machine(instance, [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] for vif in network_info: [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return self._sync_wrapper(fn, *args, **kwargs) [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self.wait() [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self[:] = self._gt.wait() [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return self._exit_event.wait() [ 793.341571] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] result = hub.switch() [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return self.greenlet.switch() [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] result = function(*args, **kwargs) [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return func(*args, **kwargs) [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] raise e [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] nwinfo = self.network_api.allocate_for_instance( [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.342143] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] created_port_ids = self._update_ports_for_instance( [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] with excutils.save_and_reraise_exception(): [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self.force_reraise() [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] raise self.value [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] updated_port = self._update_port( [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] _ensure_no_port_binding_failure(port) [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.342991] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] raise exception.PortBindingFailed(port_id=port['id']) [ 793.343384] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 793.343384] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] [ 793.343384] env[62952]: INFO nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Terminating instance [ 793.343384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.343384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquired lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.343384] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.357399] env[62952]: DEBUG nova.network.neutron [-] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.452202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "2f475040-28df-47e4-bfc4-4c843a57885f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.452202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "2f475040-28df-47e4-bfc4-4c843a57885f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.537319] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367076, 'name': ReconfigVM_Task, 'duration_secs': 0.298813} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.537319] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.537319] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3d8d333-f775-4000-92d3-166e75ce91f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.545418] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 793.545418] env[62952]: value = "task-1367077" [ 793.545418] env[62952]: _type = "Task" [ 793.545418] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.554940] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367077, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.861403] env[62952]: DEBUG nova.network.neutron [-] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.861959] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.865682] env[62952]: ERROR nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [req-d74253fa-5076-4c40-bb8d-2ea10b71e8eb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d74253fa-5076-4c40-bb8d-2ea10b71e8eb"}]}: nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 793.891380] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 793.909252] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 793.909505] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 793.932038] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 793.958858] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 793.981045] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.059245] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367077, 'name': Rename_Task, 'duration_secs': 0.165557} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.059530] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.059769] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b98cd002-f9a2-4ed1-95e4-3b4076becb73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.067052] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 794.067052] env[62952]: value = "task-1367078" [ 794.067052] env[62952]: _type = "Task" [ 794.067052] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.075544] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367078, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.303865] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17243b4-9ad0-4f77-baeb-01b40cca05a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.312128] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85ed938-c56e-4208-814c-abb71c7fb26e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.340886] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184fc3cc-c07b-4906-91f8-3854b9cbbd7b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.348151] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7ba626-7884-4e6a-ba9c-d9e085bc211d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.360855] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 794.366886] env[62952]: INFO nova.compute.manager [-] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Took 1.03 seconds to deallocate network for instance. [ 794.369067] env[62952]: DEBUG nova.compute.claims [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 794.369245] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.483357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Releasing lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.483810] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.484044] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 794.484398] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1181748c-efff-4129-bebf-b8f764dedc37 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.493835] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afeb54f1-cc74-403e-86df-5c7cf4d1da74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.517642] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65c1a63f-4645-40fe-b429-00923e60d1cb could not be found. [ 794.517885] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.518084] env[62952]: INFO nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 794.518341] env[62952]: DEBUG oslo.service.loopingcall [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.518563] env[62952]: DEBUG nova.compute.manager [-] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.518656] env[62952]: DEBUG nova.network.neutron [-] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.534288] env[62952]: DEBUG nova.network.neutron [-] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.577162] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367078, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.891821] env[62952]: DEBUG nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 76 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 794.892091] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 76 to 77 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 794.892273] env[62952]: DEBUG nova.compute.provider_tree [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.036764] env[62952]: DEBUG nova.network.neutron [-] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.061804] env[62952]: DEBUG nova.compute.manager [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Received event network-changed-5a2f6f40-4ad4-4214-b7de-c4538335eaa6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.062084] env[62952]: DEBUG nova.compute.manager [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Refreshing instance network info cache due to event network-changed-5a2f6f40-4ad4-4214-b7de-c4538335eaa6. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.062303] env[62952]: DEBUG oslo_concurrency.lockutils [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] Acquiring lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.062427] env[62952]: DEBUG oslo_concurrency.lockutils [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] Acquired lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.062607] env[62952]: DEBUG nova.network.neutron [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Refreshing network info cache for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.079658] env[62952]: DEBUG oslo_vmware.api [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367078, 'name': PowerOnVM_Task, 'duration_secs': 0.621176} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.080901] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.080901] env[62952]: INFO nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Took 6.83 seconds to spawn the instance on the hypervisor. [ 795.080901] env[62952]: DEBUG nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.081115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dea071c-ac6e-4d69-b5f9-08f284b6ee98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.397170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 4.222s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.397872] env[62952]: ERROR nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Traceback (most recent call last): [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self.driver.spawn(context, instance, image_meta, [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] vm_ref = self.build_virtual_machine(instance, [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.397872] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] for vif in network_info: [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return self._sync_wrapper(fn, *args, **kwargs) [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self.wait() [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self[:] = self._gt.wait() [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return self._exit_event.wait() [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] result = hub.switch() [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 795.398337] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return self.greenlet.switch() [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] result = function(*args, **kwargs) [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] return func(*args, **kwargs) [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] raise e [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] nwinfo = self.network_api.allocate_for_instance( [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] created_port_ids = self._update_ports_for_instance( [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] with excutils.save_and_reraise_exception(): [ 795.398689] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] self.force_reraise() [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] raise self.value [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] updated_port = self._update_port( [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] _ensure_no_port_binding_failure(port) [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] raise exception.PortBindingFailed(port_id=port['id']) [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] nova.exception.PortBindingFailed: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. [ 795.399059] env[62952]: ERROR nova.compute.manager [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] [ 795.399401] env[62952]: DEBUG nova.compute.utils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.399790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.560s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.401754] env[62952]: INFO nova.compute.claims [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.406146] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Build of instance 38e05fe2-d938-43d6-ab2b-b99fb7be51d5 was re-scheduled: Binding failed for port 8b5c79ba-cada-40ca-9d4c-a87b3783fb11, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.406704] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.406971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquiring lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.407278] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Acquired lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.407341] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.539126] env[62952]: INFO nova.compute.manager [-] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Took 1.02 seconds to deallocate network for instance. [ 795.542013] env[62952]: DEBUG nova.compute.claims [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 795.542176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.579681] env[62952]: DEBUG nova.network.neutron [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.601881] env[62952]: INFO nova.compute.manager [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Took 26.07 seconds to build instance. [ 795.684901] env[62952]: DEBUG nova.network.neutron [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.925705] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.941280] env[62952]: INFO nova.compute.manager [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Rebuilding instance [ 795.986900] env[62952]: DEBUG nova.compute.manager [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.987784] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0fb142-5bb9-4476-97d2-7f49873afbe9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.018939] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.103987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7291db4e-a4f6-44f3-a4c3-8ade46ab5d56 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "9567865d-41ae-4379-8313-e87bfc92fba8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.397s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.192218] env[62952]: DEBUG oslo_concurrency.lockutils [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] Releasing lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.192488] env[62952]: DEBUG nova.compute.manager [req-25b43baa-a922-4d08-b311-9c02b0d0a9ce req-1ba48927-7d7b-4c01-9347-1e0341c1e899 service nova] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Received event network-vif-deleted-5a2f6f40-4ad4-4214-b7de-c4538335eaa6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.497763] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.499910] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-300cf691-b01e-4bfa-8e0d-b4f271dcc9e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.506311] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 796.506311] env[62952]: value = "task-1367079" [ 796.506311] env[62952]: _type = "Task" [ 796.506311] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.515981] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.524787] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Releasing lock "refresh_cache-38e05fe2-d938-43d6-ab2b-b99fb7be51d5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.525013] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.525206] env[62952]: DEBUG nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.525412] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.542575] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.606663] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.711539] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d06ebb-8706-4b4e-b44a-5560f0ac0920 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.718602] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25396b41-081f-4766-a790-24cb8d7feac7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.746918] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445a2b46-9bbe-4ac8-b71c-d495b6113533 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.753672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9bdcf4-48a4-4691-8dd4-a0e7643ec4b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.767268] env[62952]: DEBUG nova.compute.provider_tree [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.016604] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367079, 'name': PowerOffVM_Task, 'duration_secs': 0.106892} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.016861] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.017082] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.017890] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20904c55-e1ca-4810-b923-f5a160e076df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.024580] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.024821] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c70dc173-829d-4bdd-8043-da05d7d0a1fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.045583] env[62952]: DEBUG nova.network.neutron [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.048780] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.048969] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.049155] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleting the datastore file [datastore2] 9567865d-41ae-4379-8313-e87bfc92fba8 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.049386] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eca77f0e-7601-433e-b5a4-2a87b21e14ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.055594] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 797.055594] env[62952]: value = "task-1367081" [ 797.055594] env[62952]: _type = "Task" [ 797.055594] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.063663] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367081, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.124883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.271257] env[62952]: DEBUG nova.scheduler.client.report [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.548553] env[62952]: INFO nova.compute.manager [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] [instance: 38e05fe2-d938-43d6-ab2b-b99fb7be51d5] Took 1.02 seconds to deallocate network for instance. [ 797.566557] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367081, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094552} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.566661] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.566854] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.567046] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.775803] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.776338] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.778992] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.049s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.780362] env[62952]: INFO nova.compute.claims [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.284997] env[62952]: DEBUG nova.compute.utils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.290291] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.290471] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.345933] env[62952]: DEBUG nova.policy [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '961c82479b2947618253976978b859ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9fd4a3c627a4cc9a4d60f82523789bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.586616] env[62952]: INFO nova.scheduler.client.report [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Deleted allocations for instance 38e05fe2-d938-43d6-ab2b-b99fb7be51d5 [ 798.606974] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.607227] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.607421] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.607798] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.607798] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.607897] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.608036] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.608201] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.608364] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.608572] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.608739] env[62952]: DEBUG nova.virt.hardware [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.609847] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8950c53e-2b17-48c5-a7bd-2ca15fe56e9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.618237] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a2c038-5b24-4074-9104-d22b2a368c3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.633053] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.639656] env[62952]: DEBUG oslo.service.loopingcall [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.639656] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Successfully created port: bb5a6fe2-c4da-40ad-8df3-a44c9d125c05 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.641131] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 798.641131] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5a34d50-f2e4-427d-84ff-8383ad9a05b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.657446] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.657446] env[62952]: value = "task-1367082" [ 798.657446] env[62952]: _type = "Task" [ 798.657446] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.666425] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367082, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.790893] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.097402] env[62952]: DEBUG oslo_concurrency.lockutils [None req-666a3bde-124f-4b7f-98fe-3751d4c9c438 tempest-ServersV294TestFqdnHostnames-1810021111 tempest-ServersV294TestFqdnHostnames-1810021111-project-member] Lock "38e05fe2-d938-43d6-ab2b-b99fb7be51d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.361s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.166808] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367082, 'name': CreateVM_Task, 'duration_secs': 0.253765} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.169388] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.171519] env[62952]: DEBUG oslo_vmware.service [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e66cab-4385-42a3-a444-c305152cd0e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.178415] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.178621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.179202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.179746] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e60415a-5675-46ac-ad8c-ad45a2c5dc28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.185659] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448b45ee-6024-48d5-81a4-2b1e4217c019 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.189424] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 799.189424] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52751a54-4786-068c-feb1-b953fe426bed" [ 799.189424] env[62952]: _type = "Task" [ 799.189424] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.204658] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc597376-54c5-4d6c-88fb-c999731d83a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.208138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.208435] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.208774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.208881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.209100] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.209659] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f52e866-a6b2-40d1-8fe6-6b7d6e9f4a6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.239992] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6082b7d-b954-4c2b-a1a1-00e5173d2af0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.242652] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.242873] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.243625] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0e12b8-9cb2-41a8-9161-5a4c7eda634d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.251777] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881da425-4f47-4e6e-bee7-d0b9e7749117 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.256610] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f75729a-ee8b-443c-b22b-6a474899f7d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.269324] env[62952]: DEBUG nova.compute.provider_tree [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.271853] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 799.271853] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528338c6-421e-2632-c970-54bb2d600297" [ 799.271853] env[62952]: _type = "Task" [ 799.271853] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.280950] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 799.280950] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Creating directory with path [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.280950] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18a7f764-7f4f-4668-8795-3e2e618092bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.301341] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Created directory with path [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.301341] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Fetch image to [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 799.301341] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Downloading image file data e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk on the data store datastore1 {{(pid=62952) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 799.301741] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95156ad1-cbc6-4602-b1a8-31c351edb7ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.309719] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158f217b-e971-4321-98c3-72141f40c9b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.323362] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed3203b-052e-4e4e-b0e3-538efcbd3d2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.350801] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98ec0a8-c45e-47cf-84d1-30919dc86631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.357425] env[62952]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-7eb358e3-4e44-4dad-b9f4-9daecf8e3444 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.385968] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Downloading image file data e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to the data store datastore1 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 799.449688] env[62952]: DEBUG oslo_vmware.rw_handles [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 799.604150] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.773613] env[62952]: DEBUG nova.scheduler.client.report [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.801114] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.833055] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.833055] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.833055] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.834025] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.834194] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.834425] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.834743] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.834967] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.835197] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.835763] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.835916] env[62952]: DEBUG nova.virt.hardware [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.837139] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cabd9e-916a-4699-ac23-3834035bd519 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.850907] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197f68e7-7378-436d-bb62-66cc896b457a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.868495] env[62952]: DEBUG nova.compute.manager [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Received event network-changed-bb5a6fe2-c4da-40ad-8df3-a44c9d125c05 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.868715] env[62952]: DEBUG nova.compute.manager [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Refreshing instance network info cache due to event network-changed-bb5a6fe2-c4da-40ad-8df3-a44c9d125c05. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.868949] env[62952]: DEBUG oslo_concurrency.lockutils [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] Acquiring lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.869115] env[62952]: DEBUG oslo_concurrency.lockutils [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] Acquired lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.869261] env[62952]: DEBUG nova.network.neutron [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Refreshing network info cache for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.029239] env[62952]: ERROR nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 800.029239] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.029239] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.029239] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.029239] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.029239] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.029239] env[62952]: ERROR nova.compute.manager raise self.value [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.029239] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 800.029239] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.029239] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 800.029756] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.029756] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 800.029756] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 800.029756] env[62952]: ERROR nova.compute.manager [ 800.029756] env[62952]: Traceback (most recent call last): [ 800.029756] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 800.029756] env[62952]: listener.cb(fileno) [ 800.029756] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.029756] env[62952]: result = function(*args, **kwargs) [ 800.029756] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 800.029756] env[62952]: return func(*args, **kwargs) [ 800.029756] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.029756] env[62952]: raise e [ 800.029756] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.029756] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 800.029756] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.029756] env[62952]: created_port_ids = self._update_ports_for_instance( [ 800.029756] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.029756] env[62952]: with excutils.save_and_reraise_exception(): [ 800.029756] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.029756] env[62952]: self.force_reraise() [ 800.029756] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.029756] env[62952]: raise self.value [ 800.029756] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.029756] env[62952]: updated_port = self._update_port( [ 800.029756] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.029756] env[62952]: _ensure_no_port_binding_failure(port) [ 800.029756] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.029756] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 800.030739] env[62952]: nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 800.030739] env[62952]: Removing descriptor: 15 [ 800.030739] env[62952]: ERROR nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] Traceback (most recent call last): [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] yield resources [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self.driver.spawn(context, instance, image_meta, [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.030739] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] vm_ref = self.build_virtual_machine(instance, [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] for vif in network_info: [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return self._sync_wrapper(fn, *args, **kwargs) [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self.wait() [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self[:] = self._gt.wait() [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return self._exit_event.wait() [ 800.031193] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] result = hub.switch() [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return self.greenlet.switch() [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] result = function(*args, **kwargs) [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return func(*args, **kwargs) [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] raise e [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] nwinfo = self.network_api.allocate_for_instance( [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.031635] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] created_port_ids = self._update_ports_for_instance( [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] with excutils.save_and_reraise_exception(): [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self.force_reraise() [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] raise self.value [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] updated_port = self._update_port( [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] _ensure_no_port_binding_failure(port) [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.032014] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] raise exception.PortBindingFailed(port_id=port['id']) [ 800.032405] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 800.032405] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] [ 800.032405] env[62952]: INFO nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Terminating instance [ 800.032720] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquiring lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.128536] env[62952]: DEBUG oslo_vmware.rw_handles [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 800.129343] env[62952]: DEBUG oslo_vmware.rw_handles [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 800.131658] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.268430] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Downloaded image file data e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk on the data store datastore1 {{(pid=62952) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 800.271569] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 800.271569] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Copying Virtual Disk [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk to [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.272414] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9989eb4-16eb-40f1-b92e-823d428a781a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.279444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.279866] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.284840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.478s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.286375] env[62952]: INFO nova.compute.claims [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.293017] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 800.293017] env[62952]: value = "task-1367083" [ 800.293017] env[62952]: _type = "Task" [ 800.293017] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.307036] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.402056] env[62952]: DEBUG nova.network.neutron [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.498276] env[62952]: DEBUG nova.network.neutron [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.796190] env[62952]: DEBUG nova.compute.utils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.797200] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.797403] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.811309] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367083, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.846838] env[62952]: DEBUG nova.policy [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2273fdae16cf484dbd0dc868de6faaec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c290f341462487b8ecf2c0bef2ecc00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.000677] env[62952]: DEBUG oslo_concurrency.lockutils [req-de0290df-807c-4047-8bf7-1a949b955af6 req-bbb86d57-0840-485c-b71c-3e864b9ffbd7 service nova] Releasing lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.001119] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquired lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.001319] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.151288] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Successfully created port: 3de940fa-7aef-49db-8fd2-f34b6e20e28f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.303525] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.312240] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367083, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727614} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.313191] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Copied Virtual Disk [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk to [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.313191] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleting the datastore file [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/tmp-sparse.vmdk {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.314623] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17c93272-90f0-4178-8091-900b756e2a48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.326858] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 801.326858] env[62952]: value = "task-1367084" [ 801.326858] env[62952]: _type = "Task" [ 801.326858] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.335978] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367084, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.528707] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.648157] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.759252] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defb0d00-5820-4aef-887d-e1aaa3dde520 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.767395] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64df9fbe-db91-42c2-9bba-f6c35c831127 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.797728] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b919451-e580-457d-a7d2-ee8b91f8ec60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.805195] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d371f0-deb7-4d95-9686-64f3851ea56f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.822959] env[62952]: DEBUG nova.compute.provider_tree [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.836740] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367084, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026283} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.836995] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.837353] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Moving file from [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f/e04fcbd0-b3d8-461a-890d-eda04a2a1e77 to [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77. {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 801.837550] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-908f0af6-4c4d-4691-841a-7894e75e6081 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.847172] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 801.847172] env[62952]: value = "task-1367085" [ 801.847172] env[62952]: _type = "Task" [ 801.847172] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.858481] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367085, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.894775] env[62952]: DEBUG nova.compute.manager [req-368e0cdd-4a51-4f35-9e21-89c3d3c2513d req-64a9ef53-f121-44ba-9943-ccbde02bdd14 service nova] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Received event network-vif-deleted-bb5a6fe2-c4da-40ad-8df3-a44c9d125c05 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.161035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Releasing lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.161035] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 802.161035] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 802.161035] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a18f8f1-6df8-40b7-adc4-ae5fbd3ca1c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.170460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456d70ca-87f2-49cb-b802-40a35a6ba614 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.194926] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2222309b-ce4e-46eb-a620-34553f644947 could not be found. [ 802.195351] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.197019] env[62952]: INFO nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Took 0.04 seconds to destroy the instance on the hypervisor. [ 802.197019] env[62952]: DEBUG oslo.service.loopingcall [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.197019] env[62952]: DEBUG nova.compute.manager [-] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.197019] env[62952]: DEBUG nova.network.neutron [-] [instance: 2222309b-ce4e-46eb-a620-34553f644947] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.227893] env[62952]: DEBUG nova.network.neutron [-] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.326211] env[62952]: DEBUG nova.scheduler.client.report [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.330037] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.344111] env[62952]: ERROR nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 802.344111] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.344111] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.344111] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.344111] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.344111] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.344111] env[62952]: ERROR nova.compute.manager raise self.value [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.344111] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 802.344111] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.344111] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 802.344628] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.344628] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 802.344628] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 802.344628] env[62952]: ERROR nova.compute.manager [ 802.344628] env[62952]: Traceback (most recent call last): [ 802.344628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 802.344628] env[62952]: listener.cb(fileno) [ 802.344628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.344628] env[62952]: result = function(*args, **kwargs) [ 802.344628] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.344628] env[62952]: return func(*args, **kwargs) [ 802.344628] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.344628] env[62952]: raise e [ 802.344628] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.344628] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 802.344628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.344628] env[62952]: created_port_ids = self._update_ports_for_instance( [ 802.344628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.344628] env[62952]: with excutils.save_and_reraise_exception(): [ 802.344628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.344628] env[62952]: self.force_reraise() [ 802.344628] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.344628] env[62952]: raise self.value [ 802.344628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.344628] env[62952]: updated_port = self._update_port( [ 802.344628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.344628] env[62952]: _ensure_no_port_binding_failure(port) [ 802.344628] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.344628] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 802.345508] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 802.345508] env[62952]: Removing descriptor: 15 [ 802.359863] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367085, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.037026} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.362889] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.362889] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.362889] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.363208] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.363208] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.363208] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.363208] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.363208] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.363383] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.363524] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.363710] env[62952]: DEBUG nova.virt.hardware [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.363979] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] File moved {{(pid=62952) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 802.364174] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Cleaning up location [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 802.364329] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleting the datastore file [datastore1] vmware_temp/71cfc778-539a-47b0-b150-504ad620d40f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 802.365079] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af8ed83-6d92-42c1-bb57-01d36959d2de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.367696] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fe82241-84ef-4d4e-9ee5-7bb6620b83ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.374571] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345f7725-6740-4cd2-b060-9d21b2263f2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.383258] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 802.383258] env[62952]: value = "task-1367086" [ 802.383258] env[62952]: _type = "Task" [ 802.383258] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.392026] env[62952]: ERROR nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Traceback (most recent call last): [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] yield resources [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self.driver.spawn(context, instance, image_meta, [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] vm_ref = self.build_virtual_machine(instance, [ 802.392026] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] for vif in network_info: [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] return self._sync_wrapper(fn, *args, **kwargs) [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self.wait() [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self[:] = self._gt.wait() [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] return self._exit_event.wait() [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 802.392650] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] current.throw(*self._exc) [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] result = function(*args, **kwargs) [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] return func(*args, **kwargs) [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] raise e [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] nwinfo = self.network_api.allocate_for_instance( [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] created_port_ids = self._update_ports_for_instance( [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] with excutils.save_and_reraise_exception(): [ 802.393194] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self.force_reraise() [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] raise self.value [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] updated_port = self._update_port( [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] _ensure_no_port_binding_failure(port) [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] raise exception.PortBindingFailed(port_id=port['id']) [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 802.393621] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] [ 802.393621] env[62952]: INFO nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Terminating instance [ 802.394065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquiring lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.394065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquired lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.394065] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.397678] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.733450] env[62952]: DEBUG nova.network.neutron [-] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.834250] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.834781] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.838100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.272s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.889139] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025934} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.889401] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.890120] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d608bab5-56f2-470f-862d-e585c59556f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.895286] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 802.895286] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523581d5-a776-c039-5f2d-337e08581c2b" [ 802.895286] env[62952]: _type = "Task" [ 802.895286] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.906658] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523581d5-a776-c039-5f2d-337e08581c2b, 'name': SearchDatastore_Task, 'duration_secs': 0.007956} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.906993] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.907273] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.907518] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f70c3176-7c03-4ba8-989b-993578e472bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.913606] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 802.913606] env[62952]: value = "task-1367087" [ 802.913606] env[62952]: _type = "Task" [ 802.913606] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.917372] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.922934] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367087, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.001586] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.237422] env[62952]: INFO nova.compute.manager [-] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Took 1.04 seconds to deallocate network for instance. [ 803.240852] env[62952]: DEBUG nova.compute.claims [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 803.241144] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.342330] env[62952]: DEBUG nova.compute.utils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.347348] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.347525] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.404871] env[62952]: DEBUG nova.policy [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21dd2d81b6ee48e2a22cc21e125e0f47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a76e4e4643c84e04ac632efaecdd652f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.422686] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367087, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495015} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.425260] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.425693] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.426762] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3274b8e-cd56-41a9-ac17-4a1802daad06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.434230] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 803.434230] env[62952]: value = "task-1367088" [ 803.434230] env[62952]: _type = "Task" [ 803.434230] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.451261] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367088, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.504826] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Releasing lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.506164] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 803.506362] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.506646] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51df0a27-9a99-4aca-a65d-1d28601e9645 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.518512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0769b5cb-f0dd-490d-9f62-c8bdcc880f96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.546862] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 74361d48-0a82-453a-a4a1-45173c49e901 could not be found. [ 803.547102] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.547285] env[62952]: INFO nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Took 0.04 seconds to destroy the instance on the hypervisor. [ 803.547526] env[62952]: DEBUG oslo.service.loopingcall [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.550576] env[62952]: DEBUG nova.compute.manager [-] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.550700] env[62952]: DEBUG nova.network.neutron [-] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 803.570684] env[62952]: DEBUG nova.network.neutron [-] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.745340] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Successfully created port: d78daeab-1f1d-441c-8b47-dea2fbe31ad4 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.765514] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dbd4a9e-509a-45fd-8cf1-257de9b83748 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.773023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a939990-cc25-41f9-95f8-07e1b4e57225 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.808184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9819931-5294-4ebd-9dc6-c7ed163c4774 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.815672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81d864c-2c06-4a13-a8e3-431eb483ba07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.829998] env[62952]: DEBUG nova.compute.provider_tree [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.850469] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.944452] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367088, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063768} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.944764] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.945636] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5345ff7c-800e-4090-bd17-581951092eda {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.971026] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.971026] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-592a0867-b169-4c8a-8526-959fcf34c82e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.990840] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 803.990840] env[62952]: value = "task-1367089" [ 803.990840] env[62952]: _type = "Task" [ 803.990840] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.997891] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367089, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.053823] env[62952]: DEBUG nova.compute.manager [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Received event network-changed-3de940fa-7aef-49db-8fd2-f34b6e20e28f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.054068] env[62952]: DEBUG nova.compute.manager [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Refreshing instance network info cache due to event network-changed-3de940fa-7aef-49db-8fd2-f34b6e20e28f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.054294] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] Acquiring lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.054438] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] Acquired lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.054749] env[62952]: DEBUG nova.network.neutron [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Refreshing network info cache for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.072552] env[62952]: DEBUG nova.network.neutron [-] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.335728] env[62952]: DEBUG nova.scheduler.client.report [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.502631] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367089, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.576255] env[62952]: INFO nova.compute.manager [-] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Took 1.02 seconds to deallocate network for instance. [ 804.578919] env[62952]: DEBUG nova.compute.claims [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 804.579149] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.590043] env[62952]: DEBUG nova.network.neutron [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.783786] env[62952]: DEBUG nova.network.neutron [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.841615] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.842169] env[62952]: ERROR nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Traceback (most recent call last): [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self.driver.spawn(context, instance, image_meta, [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] vm_ref = self.build_virtual_machine(instance, [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.842169] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] for vif in network_info: [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return self._sync_wrapper(fn, *args, **kwargs) [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self.wait() [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self[:] = self._gt.wait() [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return self._exit_event.wait() [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] result = hub.switch() [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 804.842742] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return self.greenlet.switch() [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] result = function(*args, **kwargs) [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] return func(*args, **kwargs) [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] raise e [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] nwinfo = self.network_api.allocate_for_instance( [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] created_port_ids = self._update_ports_for_instance( [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] with excutils.save_and_reraise_exception(): [ 804.843425] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] self.force_reraise() [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] raise self.value [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] updated_port = self._update_port( [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] _ensure_no_port_binding_failure(port) [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] raise exception.PortBindingFailed(port_id=port['id']) [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] nova.exception.PortBindingFailed: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. [ 804.843927] env[62952]: ERROR nova.compute.manager [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] [ 804.844245] env[62952]: DEBUG nova.compute.utils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 804.849026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.815s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.849026] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Build of instance 5283bf79-bf20-4a38-912f-ced401fbd691 was re-scheduled: Binding failed for port 937f84a5-94cd-42da-a5de-53578453551b, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 804.849322] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 804.849322] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquiring lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.849434] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Acquired lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.849578] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.866459] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.898066] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.898066] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.898066] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.898066] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.898303] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.898303] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.898303] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.898496] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.898664] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.899558] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.899725] env[62952]: DEBUG nova.virt.hardware [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.901436] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25bccd05-5094-4159-8152-eee81b5fe28d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.913107] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3119fc6f-79eb-49e2-8a73-5015596eb28e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.991812] env[62952]: ERROR nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 804.991812] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.991812] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.991812] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.991812] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.991812] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.991812] env[62952]: ERROR nova.compute.manager raise self.value [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.991812] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 804.991812] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.991812] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 804.992892] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.992892] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 804.992892] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 804.992892] env[62952]: ERROR nova.compute.manager [ 804.992892] env[62952]: Traceback (most recent call last): [ 804.992892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 804.992892] env[62952]: listener.cb(fileno) [ 804.992892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.992892] env[62952]: result = function(*args, **kwargs) [ 804.992892] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.992892] env[62952]: return func(*args, **kwargs) [ 804.992892] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.992892] env[62952]: raise e [ 804.992892] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.992892] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 804.992892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.992892] env[62952]: created_port_ids = self._update_ports_for_instance( [ 804.992892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.992892] env[62952]: with excutils.save_and_reraise_exception(): [ 804.992892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.992892] env[62952]: self.force_reraise() [ 804.992892] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.992892] env[62952]: raise self.value [ 804.992892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.992892] env[62952]: updated_port = self._update_port( [ 804.992892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.992892] env[62952]: _ensure_no_port_binding_failure(port) [ 804.992892] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.992892] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 804.996730] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 804.996730] env[62952]: Removing descriptor: 15 [ 804.996730] env[62952]: ERROR nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Traceback (most recent call last): [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] yield resources [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self.driver.spawn(context, instance, image_meta, [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.996730] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] vm_ref = self.build_virtual_machine(instance, [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] for vif in network_info: [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return self._sync_wrapper(fn, *args, **kwargs) [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self.wait() [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self[:] = self._gt.wait() [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return self._exit_event.wait() [ 804.997327] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] result = hub.switch() [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return self.greenlet.switch() [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] result = function(*args, **kwargs) [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return func(*args, **kwargs) [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] raise e [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] nwinfo = self.network_api.allocate_for_instance( [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.997681] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] created_port_ids = self._update_ports_for_instance( [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] with excutils.save_and_reraise_exception(): [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self.force_reraise() [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] raise self.value [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] updated_port = self._update_port( [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] _ensure_no_port_binding_failure(port) [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.998102] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] raise exception.PortBindingFailed(port_id=port['id']) [ 804.998429] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 804.998429] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] [ 804.998429] env[62952]: INFO nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Terminating instance [ 804.998429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquiring lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.998554] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquired lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.998676] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.005229] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367089, 'name': ReconfigVM_Task, 'duration_secs': 0.694983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.005471] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 9567865d-41ae-4379-8313-e87bfc92fba8/9567865d-41ae-4379-8313-e87bfc92fba8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.006078] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20d2f262-a30d-4e57-9731-75386b8642b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.013353] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 805.013353] env[62952]: value = "task-1367090" [ 805.013353] env[62952]: _type = "Task" [ 805.013353] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.021815] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367090, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.286424] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] Releasing lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.286695] env[62952]: DEBUG nova.compute.manager [req-f9286045-7154-4ade-b110-9a0794d46b17 req-47322f9e-a3a3-408d-ab2a-46680bcafa4a service nova] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Received event network-vif-deleted-3de940fa-7aef-49db-8fd2-f34b6e20e28f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.372143] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.496548] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.529712] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367090, 'name': Rename_Task, 'duration_secs': 0.149278} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.530924] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.540641] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.541152] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bbaf800-268d-4a4e-b817-b2ac9761ae7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.549840] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 805.549840] env[62952]: value = "task-1367091" [ 805.549840] env[62952]: _type = "Task" [ 805.549840] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.566336] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367091, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.626499] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.765633] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bf1ae2-1049-46d8-b2d5-10d4ad96e47a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.774481] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4cf550-2607-4bb1-ae9f-45358f8f6ac6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.810053] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d24aa8e-f6a3-4494-8380-f60ea27ad4e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.818041] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83276d5f-089f-4da3-b7d3-aea79b5caf47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.835127] env[62952]: DEBUG nova.compute.provider_tree [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.998840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Releasing lock "refresh_cache-5283bf79-bf20-4a38-912f-ced401fbd691" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.999293] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 805.999293] env[62952]: DEBUG nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.001179] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.017685] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.059650] env[62952]: DEBUG oslo_vmware.api [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367091, 'name': PowerOnVM_Task, 'duration_secs': 0.418178} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.059919] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.060169] env[62952]: DEBUG nova.compute.manager [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.061031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82081e3-dce9-4c6d-919b-2cb78dbc6e3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.099569] env[62952]: DEBUG nova.compute.manager [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Received event network-changed-d78daeab-1f1d-441c-8b47-dea2fbe31ad4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.099569] env[62952]: DEBUG nova.compute.manager [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Refreshing instance network info cache due to event network-changed-d78daeab-1f1d-441c-8b47-dea2fbe31ad4. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.099569] env[62952]: DEBUG oslo_concurrency.lockutils [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] Acquiring lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.134021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Releasing lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.134021] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.134021] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.134021] env[62952]: DEBUG oslo_concurrency.lockutils [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] Acquired lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.134021] env[62952]: DEBUG nova.network.neutron [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Refreshing network info cache for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.134626] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da290a80-7846-43bc-889d-925368373d25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.144504] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a88512b-9c57-4179-8aa5-a12160c1de2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.171510] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91660994-feca-40d8-8b75-d3c79110a19e could not be found. [ 806.171901] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.172220] env[62952]: INFO nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 806.172586] env[62952]: DEBUG oslo.service.loopingcall [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.172892] env[62952]: DEBUG nova.compute.manager [-] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.173100] env[62952]: DEBUG nova.network.neutron [-] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.194385] env[62952]: DEBUG nova.network.neutron [-] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.336277] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "aef28168-98a7-4f65-80e7-731633339abf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.336277] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "aef28168-98a7-4f65-80e7-731633339abf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.339794] env[62952]: DEBUG nova.scheduler.client.report [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.519874] env[62952]: DEBUG nova.network.neutron [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.577558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.653712] env[62952]: DEBUG nova.network.neutron [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.698342] env[62952]: DEBUG nova.network.neutron [-] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.745902] env[62952]: DEBUG nova.network.neutron [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.843595] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "9567865d-41ae-4379-8313-e87bfc92fba8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.843801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "9567865d-41ae-4379-8313-e87bfc92fba8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.844050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "9567865d-41ae-4379-8313-e87bfc92fba8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.844243] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "9567865d-41ae-4379-8313-e87bfc92fba8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.844413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "9567865d-41ae-4379-8313-e87bfc92fba8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.846520] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.847476] env[62952]: ERROR nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Traceback (most recent call last): [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self.driver.spawn(context, instance, image_meta, [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] vm_ref = self.build_virtual_machine(instance, [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.847476] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] for vif in network_info: [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return self._sync_wrapper(fn, *args, **kwargs) [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self.wait() [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self[:] = self._gt.wait() [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return self._exit_event.wait() [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] result = hub.switch() [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 806.847867] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return self.greenlet.switch() [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] result = function(*args, **kwargs) [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] return func(*args, **kwargs) [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] raise e [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] nwinfo = self.network_api.allocate_for_instance( [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] created_port_ids = self._update_ports_for_instance( [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] with excutils.save_and_reraise_exception(): [ 806.848293] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] self.force_reraise() [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] raise self.value [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] updated_port = self._update_port( [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] _ensure_no_port_binding_failure(port) [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] raise exception.PortBindingFailed(port_id=port['id']) [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] nova.exception.PortBindingFailed: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. [ 806.848702] env[62952]: ERROR nova.compute.manager [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] [ 806.849108] env[62952]: DEBUG nova.compute.utils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 806.849596] env[62952]: INFO nova.compute.manager [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Terminating instance [ 806.850992] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.188s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.852388] env[62952]: INFO nova.compute.claims [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.854788] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "refresh_cache-9567865d-41ae-4379-8313-e87bfc92fba8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.854943] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquired lock "refresh_cache-9567865d-41ae-4379-8313-e87bfc92fba8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.855116] env[62952]: DEBUG nova.network.neutron [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.864510] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Build of instance fc5b891b-8432-4b2f-97d9-c93ca0a18783 was re-scheduled: Binding failed for port 272b9bc4-bacd-48d8-b805-6a6ce692bafe, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 806.864974] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 806.865202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquiring lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.865355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Acquired lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.865514] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.023466] env[62952]: INFO nova.compute.manager [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] [instance: 5283bf79-bf20-4a38-912f-ced401fbd691] Took 1.02 seconds to deallocate network for instance. [ 807.203089] env[62952]: INFO nova.compute.manager [-] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Took 1.03 seconds to deallocate network for instance. [ 807.207696] env[62952]: DEBUG nova.compute.claims [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 807.207873] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.248616] env[62952]: DEBUG oslo_concurrency.lockutils [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] Releasing lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.248861] env[62952]: DEBUG nova.compute.manager [req-8f7556c1-8e8c-48a0-8030-a8c9f7f30630 req-990d94c9-7926-4272-88d4-e18e917c00f9 service nova] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Received event network-vif-deleted-d78daeab-1f1d-441c-8b47-dea2fbe31ad4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.374860] env[62952]: DEBUG nova.network.neutron [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.388746] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.462567] env[62952]: DEBUG nova.network.neutron [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.499943] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.968757] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Releasing lock "refresh_cache-9567865d-41ae-4379-8313-e87bfc92fba8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.969463] env[62952]: DEBUG nova.compute.manager [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.969658] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.970927] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16f0330-1974-4f1e-aa62-b18dd56095af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.978630] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.981019] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df194eaa-0509-45a6-823e-395d6778ccbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.988704] env[62952]: DEBUG oslo_vmware.api [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 807.988704] env[62952]: value = "task-1367092" [ 807.988704] env[62952]: _type = "Task" [ 807.988704] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.999815] env[62952]: DEBUG oslo_vmware.api [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.002650] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Releasing lock "refresh_cache-fc5b891b-8432-4b2f-97d9-c93ca0a18783" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.002650] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.002790] env[62952]: DEBUG nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.003052] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.021177] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.051619] env[62952]: INFO nova.scheduler.client.report [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Deleted allocations for instance 5283bf79-bf20-4a38-912f-ced401fbd691 [ 808.197914] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1170afdb-0383-4d97-ab22-f56d618b71ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.205292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b24daf-1e55-4877-8e3e-a0569e4ce297 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.234971] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a87e92c-ed5d-4363-9f6e-4e99b4193e94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.242385] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772d4b39-3060-45be-9099-b8612ff1f369 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.255613] env[62952]: DEBUG nova.compute.provider_tree [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.497929] env[62952]: DEBUG oslo_vmware.api [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367092, 'name': PowerOffVM_Task, 'duration_secs': 0.193928} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.498624] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.498624] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.498822] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca161f71-a4c5-4448-b4b5-0895446c94fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.524373] env[62952]: DEBUG nova.network.neutron [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.525502] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.525687] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.525860] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleting the datastore file [datastore1] 9567865d-41ae-4379-8313-e87bfc92fba8 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.526318] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a60df37e-5c05-4a64-b95c-44469dc79de3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.532362] env[62952]: DEBUG oslo_vmware.api [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for the task: (returnval){ [ 808.532362] env[62952]: value = "task-1367094" [ 808.532362] env[62952]: _type = "Task" [ 808.532362] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.540736] env[62952]: DEBUG oslo_vmware.api [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.560944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6bc5a26-38b6-45b6-ba90-3318437b43d1 tempest-ServersTestMultiNic-1870011473 tempest-ServersTestMultiNic-1870011473-project-member] Lock "5283bf79-bf20-4a38-912f-ced401fbd691" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.659s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.760404] env[62952]: DEBUG nova.scheduler.client.report [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.027488] env[62952]: INFO nova.compute.manager [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] [instance: fc5b891b-8432-4b2f-97d9-c93ca0a18783] Took 1.02 seconds to deallocate network for instance. [ 809.043207] env[62952]: DEBUG oslo_vmware.api [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Task: {'id': task-1367094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115624} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.043459] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.043637] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 809.043828] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.044054] env[62952]: INFO nova.compute.manager [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Took 1.07 seconds to destroy the instance on the hypervisor. [ 809.044305] env[62952]: DEBUG oslo.service.loopingcall [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.044496] env[62952]: DEBUG nova.compute.manager [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.044589] env[62952]: DEBUG nova.network.neutron [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.063292] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.070346] env[62952]: DEBUG nova.network.neutron [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.264500] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.265086] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.267776] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.018s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.577512] env[62952]: DEBUG nova.network.neutron [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.594443] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.774053] env[62952]: DEBUG nova.compute.utils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.779673] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.779858] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 809.859129] env[62952]: DEBUG nova.policy [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd942d80343945d5b9730a96d884f22b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f550284230694a76afb9ffd94a7f9175', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.068436] env[62952]: INFO nova.scheduler.client.report [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Deleted allocations for instance fc5b891b-8432-4b2f-97d9-c93ca0a18783 [ 810.081240] env[62952]: INFO nova.compute.manager [-] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Took 1.04 seconds to deallocate network for instance. [ 810.285326] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.309996] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 9567865d-41ae-4379-8313-e87bfc92fba8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.310179] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 0f8162d7-c5e6-4c66-8a10-bb06998738b5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.310307] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 65c1a63f-4645-40fe-b429-00923e60d1cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.310429] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 2222309b-ce4e-46eb-a620-34553f644947 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.310547] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 74361d48-0a82-453a-a4a1-45173c49e901 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.310663] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 91660994-feca-40d8-8b75-d3c79110a19e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.310782] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance c7771b47-cbe8-4696-9a21-9204501c8877 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 810.576655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1dd95b8a-3d77-47a3-b1fe-9aa38d085e39 tempest-ServerMetadataTestJSON-814045322 tempest-ServerMetadataTestJSON-814045322-project-member] Lock "fc5b891b-8432-4b2f-97d9-c93ca0a18783" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.032s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.587206] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Successfully created port: fa638cdd-dbbc-466f-9d99-c3a874f1ca12 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.590439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.815498] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4be2f0a4-5f57-4a80-9450-6d61d344799e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 811.079980] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.293183] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.319953] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 6f085d0c-c440-4bba-9bf8-cf2f70742748 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 811.329934] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.329934] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.329934] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.330228] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.330228] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.330556] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.330886] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.330958] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.331153] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.331312] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.331478] env[62952]: DEBUG nova.virt.hardware [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.333016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc23b1a-f23f-45fd-b2b6-9333b9bae241 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.347592] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c254f237-d34d-4e37-a1b0-e100d6163865 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.611121] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.825928] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 95c07667-e9d8-4cc3-b135-f80d196d179f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 812.064391] env[62952]: DEBUG nova.compute.manager [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Received event network-changed-fa638cdd-dbbc-466f-9d99-c3a874f1ca12 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.064589] env[62952]: DEBUG nova.compute.manager [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Refreshing instance network info cache due to event network-changed-fa638cdd-dbbc-466f-9d99-c3a874f1ca12. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.064799] env[62952]: DEBUG oslo_concurrency.lockutils [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] Acquiring lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.064947] env[62952]: DEBUG oslo_concurrency.lockutils [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] Acquired lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.065119] env[62952]: DEBUG nova.network.neutron [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Refreshing network info cache for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.224124] env[62952]: ERROR nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 812.224124] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.224124] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.224124] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.224124] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.224124] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.224124] env[62952]: ERROR nova.compute.manager raise self.value [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.224124] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 812.224124] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.224124] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 812.224956] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.224956] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 812.224956] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 812.224956] env[62952]: ERROR nova.compute.manager [ 812.224956] env[62952]: Traceback (most recent call last): [ 812.224956] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 812.224956] env[62952]: listener.cb(fileno) [ 812.224956] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.224956] env[62952]: result = function(*args, **kwargs) [ 812.224956] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 812.224956] env[62952]: return func(*args, **kwargs) [ 812.224956] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 812.224956] env[62952]: raise e [ 812.224956] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.224956] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 812.224956] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.224956] env[62952]: created_port_ids = self._update_ports_for_instance( [ 812.224956] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.224956] env[62952]: with excutils.save_and_reraise_exception(): [ 812.224956] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.224956] env[62952]: self.force_reraise() [ 812.224956] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.224956] env[62952]: raise self.value [ 812.224956] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.224956] env[62952]: updated_port = self._update_port( [ 812.224956] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.224956] env[62952]: _ensure_no_port_binding_failure(port) [ 812.224956] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.224956] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 812.225713] env[62952]: nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 812.225713] env[62952]: Removing descriptor: 19 [ 812.225713] env[62952]: ERROR nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Traceback (most recent call last): [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] yield resources [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self.driver.spawn(context, instance, image_meta, [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self._vmops.spawn(context, instance, image_meta, injected_files, [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 812.225713] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] vm_ref = self.build_virtual_machine(instance, [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] vif_infos = vmwarevif.get_vif_info(self._session, [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] for vif in network_info: [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return self._sync_wrapper(fn, *args, **kwargs) [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self.wait() [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self[:] = self._gt.wait() [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return self._exit_event.wait() [ 812.226604] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] result = hub.switch() [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return self.greenlet.switch() [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] result = function(*args, **kwargs) [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return func(*args, **kwargs) [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] raise e [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] nwinfo = self.network_api.allocate_for_instance( [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.227048] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] created_port_ids = self._update_ports_for_instance( [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] with excutils.save_and_reraise_exception(): [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self.force_reraise() [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] raise self.value [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] updated_port = self._update_port( [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] _ensure_no_port_binding_failure(port) [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.227446] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] raise exception.PortBindingFailed(port_id=port['id']) [ 812.227845] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 812.227845] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] [ 812.227845] env[62952]: INFO nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Terminating instance [ 812.229179] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquiring lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.332323] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ca7fc13d-e9be-481c-ae58-78377627800c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 812.589797] env[62952]: DEBUG nova.network.neutron [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.710260] env[62952]: DEBUG nova.network.neutron [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.838445] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 34b6c13c-c95b-4cc2-b6eb-cc9374535aec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 813.213039] env[62952]: DEBUG oslo_concurrency.lockutils [req-9ab72cca-0459-4e74-b7b5-95f1ad2f867a req-ba4d1f34-3cee-4d91-9dfc-09ee974e6b3c service nova] Releasing lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.213553] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquired lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.213757] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.342023] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4846f1d8-7628-405e-bef0-0d795a973a34 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 813.731542] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.815908] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.845601] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.845601] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.845601] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 16224e22-7091-4381-9f40-3f52f24bb724 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 814.111376] env[62952]: DEBUG nova.compute.manager [req-8b8f787b-b171-4657-ad3e-112ca960acc1 req-e708fc04-93fc-4a5c-be49-80d095f10818 service nova] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Received event network-vif-deleted-fa638cdd-dbbc-466f-9d99-c3a874f1ca12 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.319116] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Releasing lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.319532] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.321872] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 814.321872] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5225f2fe-5259-465b-9e37-ddfdabf9297d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.329506] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd776f8-bffa-4160-81be-d761b5d0325d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.350597] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance b4ea4cff-ada0-4355-a514-d8b179616535 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 814.353034] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c7771b47-cbe8-4696-9a21-9204501c8877 could not be found. [ 814.353034] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.353034] env[62952]: INFO nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Took 0.03 seconds to destroy the instance on the hypervisor. [ 814.353255] env[62952]: DEBUG oslo.service.loopingcall [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.353385] env[62952]: DEBUG nova.compute.manager [-] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 814.353480] env[62952]: DEBUG nova.network.neutron [-] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 814.369716] env[62952]: DEBUG nova.network.neutron [-] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.853796] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 39948315-6438-48f1-883b-b376e0786650 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 814.872511] env[62952]: DEBUG nova.network.neutron [-] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.355891] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 69c93772-822a-4f5b-b2d1-95e82a46f286 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 815.375175] env[62952]: INFO nova.compute.manager [-] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Took 1.02 seconds to deallocate network for instance. [ 815.377423] env[62952]: DEBUG nova.compute.claims [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 815.377592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.860058] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 62d47ebb-8782-4062-a27e-556c758ed475 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 816.362945] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 56aa8186-2bcb-47fc-a832-31654cbf218b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 816.866250] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 351b1672-ddb3-4562-9965-d65a7d869e82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 817.368789] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance d2faf1e0-cfd1-4d87-ba77-0af92dc16643 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 817.871799] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a63ec2ac-4484-4360-962b-105a5338c1b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 818.374755] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 802fd770-38aa-4417-b25e-fd83510bceef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 818.877252] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 2f475040-28df-47e4-bfc4-4c843a57885f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 819.381310] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 819.381692] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 819.381749] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 819.692251] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1497bc99-d1c4-4484-941e-d044e14ca5ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.700300] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62036458-7658-4499-b3e2-f9b058fda4f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.729723] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a776b1ca-4d2c-4a1f-9d4a-52c84e8350fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.736461] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be45be62-aa63-4755-9ca5-ee05265cc0f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.749950] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.253503] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.758880] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 820.760077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.491s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.760450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.391s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.554045] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd062e1b-158f-4c50-bbf3-67854098a287 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.564019] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d25a8e0-8c32-48ce-b082-41bb06c4d7a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.593902] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676a457b-3969-46c8-80d4-4a51508b8aac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.601433] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2d1fd6-f94c-4f2c-b8d4-15763855e757 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.614219] env[62952]: DEBUG nova.compute.provider_tree [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.117512] env[62952]: DEBUG nova.scheduler.client.report [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.622664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.862s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.624023] env[62952]: ERROR nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Traceback (most recent call last): [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self.driver.spawn(context, instance, image_meta, [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] vm_ref = self.build_virtual_machine(instance, [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 822.624023] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] for vif in network_info: [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return self._sync_wrapper(fn, *args, **kwargs) [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self.wait() [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self[:] = self._gt.wait() [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return self._exit_event.wait() [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] result = hub.switch() [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 822.624496] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return self.greenlet.switch() [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] result = function(*args, **kwargs) [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] return func(*args, **kwargs) [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] raise e [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] nwinfo = self.network_api.allocate_for_instance( [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] created_port_ids = self._update_ports_for_instance( [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] with excutils.save_and_reraise_exception(): [ 822.625078] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] self.force_reraise() [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] raise self.value [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] updated_port = self._update_port( [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] _ensure_no_port_binding_failure(port) [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] raise exception.PortBindingFailed(port_id=port['id']) [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] nova.exception.PortBindingFailed: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. [ 822.625501] env[62952]: ERROR nova.compute.manager [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] [ 822.625866] env[62952]: DEBUG nova.compute.utils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 822.626513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.084s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.632048] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Build of instance 0f8162d7-c5e6-4c66-8a10-bb06998738b5 was re-scheduled: Binding failed for port 3af034e9-59a9-4803-8390-ec339b7ae0d5, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 822.632356] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 822.632696] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.632949] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquired lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.633229] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.157755] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.248828] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.484961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1eacca-131b-490e-91cd-84fefaca7e89 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.493250] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05109bc-8e2d-4abd-ac1e-93b75605f0fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.525240] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f997d1b-c54a-412b-ba1c-dbf6e4116b4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.532871] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b069576e-4674-485e-b587-7b71e1db9fc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.547051] env[62952]: DEBUG nova.compute.provider_tree [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.580765] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.580981] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.754148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Releasing lock "refresh_cache-0f8162d7-c5e6-4c66-8a10-bb06998738b5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.754370] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 823.754571] env[62952]: DEBUG nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.754735] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.770819] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.049625] env[62952]: DEBUG nova.scheduler.client.report [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.086177] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 824.087038] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 824.087038] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 824.273437] env[62952]: DEBUG nova.network.neutron [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.554026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.554618] env[62952]: ERROR nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Traceback (most recent call last): [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self.driver.spawn(context, instance, image_meta, [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] vm_ref = self.build_virtual_machine(instance, [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 824.554618] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] for vif in network_info: [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return self._sync_wrapper(fn, *args, **kwargs) [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self.wait() [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self[:] = self._gt.wait() [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return self._exit_event.wait() [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] result = hub.switch() [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 824.555723] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return self.greenlet.switch() [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] result = function(*args, **kwargs) [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] return func(*args, **kwargs) [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] raise e [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] nwinfo = self.network_api.allocate_for_instance( [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] created_port_ids = self._update_ports_for_instance( [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] with excutils.save_and_reraise_exception(): [ 824.557262] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] self.force_reraise() [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] raise self.value [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] updated_port = self._update_port( [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] _ensure_no_port_binding_failure(port) [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] raise exception.PortBindingFailed(port_id=port['id']) [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] nova.exception.PortBindingFailed: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. [ 824.557840] env[62952]: ERROR nova.compute.manager [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] [ 824.558327] env[62952]: DEBUG nova.compute.utils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 824.558327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.432s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.558327] env[62952]: INFO nova.compute.claims [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.560802] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Build of instance 65c1a63f-4645-40fe-b429-00923e60d1cb was re-scheduled: Binding failed for port 5a2f6f40-4ad4-4214-b7de-c4538335eaa6, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 824.561295] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 824.561517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquiring lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.561660] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Acquired lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.561816] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.591332] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.591505] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.591639] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.591766] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.617742] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-9567865d-41ae-4379-8313-e87bfc92fba8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.617893] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-9567865d-41ae-4379-8313-e87bfc92fba8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.618081] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 824.618277] env[62952]: DEBUG nova.objects.instance [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lazy-loading 'info_cache' on Instance uuid 9567865d-41ae-4379-8313-e87bfc92fba8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.776520] env[62952]: INFO nova.compute.manager [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 0f8162d7-c5e6-4c66-8a10-bb06998738b5] Took 1.02 seconds to deallocate network for instance. [ 825.081802] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.188307] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.637676] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.692219] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Releasing lock "refresh_cache-65c1a63f-4645-40fe-b429-00923e60d1cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.692444] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 825.692620] env[62952]: DEBUG nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.692787] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.708471] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.802798] env[62952]: INFO nova.scheduler.client.report [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Deleted allocations for instance 0f8162d7-c5e6-4c66-8a10-bb06998738b5 [ 825.899418] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c297f2-611b-4350-b480-dd79c8aeba79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.905764] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb7fe46-e388-4b2d-b58d-ffe0a474f736 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.941018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a233b240-b1c1-4e5f-aad5-4ba7ee6eb6b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.950171] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c801f0e-5ad8-4c71-b487-a3435667b42d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.965451] env[62952]: DEBUG nova.compute.provider_tree [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.195596] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.210763] env[62952]: DEBUG nova.network.neutron [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.312537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-452fe707-e83c-4fef-8a40-350f6c5d33cc tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "0f8162d7-c5e6-4c66-8a10-bb06998738b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.403s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.469166] env[62952]: DEBUG nova.scheduler.client.report [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.698847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-9567865d-41ae-4379-8313-e87bfc92fba8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.698847] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 826.699182] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.699182] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.699304] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.699509] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.699705] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.699888] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.700057] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 826.700220] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.712542] env[62952]: INFO nova.compute.manager [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] [instance: 65c1a63f-4645-40fe-b429-00923e60d1cb] Took 1.02 seconds to deallocate network for instance. [ 826.815527] env[62952]: DEBUG nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.973869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.974435] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.976953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.845s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.978543] env[62952]: INFO nova.compute.claims [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.203011] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.333422] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.483746] env[62952]: DEBUG nova.compute.utils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.487280] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.487280] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.526288] env[62952]: DEBUG nova.policy [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55ffb58a3b3a4ad29699d48ecd6c0700', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d564dc3c75e430dbaf1f90a9c90c18b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.753602] env[62952]: INFO nova.scheduler.client.report [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Deleted allocations for instance 65c1a63f-4645-40fe-b429-00923e60d1cb [ 827.851413] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Successfully created port: 4e8bbee8-553c-404c-a2a3-eaefff21a6f9 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.987739] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.263388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8df1282d-2f7a-4089-806d-187ccd308ead tempest-ServerRescueNegativeTestJSON-785829795 tempest-ServerRescueNegativeTestJSON-785829795-project-member] Lock "65c1a63f-4645-40fe-b429-00923e60d1cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.367s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.318026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c18e0c-2d13-42a9-b59f-bdf26265b0d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.329888] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545e2a3a-fa3d-4e64-863f-ffdf2880c80d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.359983] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f506d94-acc4-4605-9598-703b5e80fd51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.368430] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0195d6ca-41d2-4fad-95c5-c3c40981ebe7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.381815] env[62952]: DEBUG nova.compute.provider_tree [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.626063] env[62952]: DEBUG nova.compute.manager [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Received event network-changed-4e8bbee8-553c-404c-a2a3-eaefff21a6f9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.626593] env[62952]: DEBUG nova.compute.manager [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Refreshing instance network info cache due to event network-changed-4e8bbee8-553c-404c-a2a3-eaefff21a6f9. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.626593] env[62952]: DEBUG oslo_concurrency.lockutils [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] Acquiring lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.626593] env[62952]: DEBUG oslo_concurrency.lockutils [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] Acquired lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.626797] env[62952]: DEBUG nova.network.neutron [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Refreshing network info cache for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.765861] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.797998] env[62952]: ERROR nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 828.797998] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.797998] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.797998] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.797998] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.797998] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.797998] env[62952]: ERROR nova.compute.manager raise self.value [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.797998] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 828.797998] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.797998] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 828.798615] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.798615] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 828.798615] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 828.798615] env[62952]: ERROR nova.compute.manager [ 828.798615] env[62952]: Traceback (most recent call last): [ 828.798615] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 828.798615] env[62952]: listener.cb(fileno) [ 828.798615] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.798615] env[62952]: result = function(*args, **kwargs) [ 828.798615] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.798615] env[62952]: return func(*args, **kwargs) [ 828.798615] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.798615] env[62952]: raise e [ 828.798615] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.798615] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 828.798615] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.798615] env[62952]: created_port_ids = self._update_ports_for_instance( [ 828.798615] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.798615] env[62952]: with excutils.save_and_reraise_exception(): [ 828.798615] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.798615] env[62952]: self.force_reraise() [ 828.798615] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.798615] env[62952]: raise self.value [ 828.798615] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.798615] env[62952]: updated_port = self._update_port( [ 828.798615] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.798615] env[62952]: _ensure_no_port_binding_failure(port) [ 828.798615] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.798615] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 828.799616] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 828.799616] env[62952]: Removing descriptor: 19 [ 828.886070] env[62952]: DEBUG nova.scheduler.client.report [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.999121] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.024135] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.024378] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.024536] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.024713] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.024856] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.025010] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.025232] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.025419] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.025592] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.025751] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.025922] env[62952]: DEBUG nova.virt.hardware [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.026797] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303ed01f-318b-4195-adef-5bfce567ffa9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.036313] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5b3e9d-def3-4e17-94bc-5d1fe50a999d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.049592] env[62952]: ERROR nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Traceback (most recent call last): [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] yield resources [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self.driver.spawn(context, instance, image_meta, [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] vm_ref = self.build_virtual_machine(instance, [ 829.049592] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] vif_infos = vmwarevif.get_vif_info(self._session, [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] for vif in network_info: [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] return self._sync_wrapper(fn, *args, **kwargs) [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self.wait() [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self[:] = self._gt.wait() [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] return self._exit_event.wait() [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 829.050060] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] current.throw(*self._exc) [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] result = function(*args, **kwargs) [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] return func(*args, **kwargs) [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] raise e [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] nwinfo = self.network_api.allocate_for_instance( [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] created_port_ids = self._update_ports_for_instance( [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] with excutils.save_and_reraise_exception(): [ 829.050477] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self.force_reraise() [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] raise self.value [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] updated_port = self._update_port( [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] _ensure_no_port_binding_failure(port) [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] raise exception.PortBindingFailed(port_id=port['id']) [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 829.050909] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] [ 829.050909] env[62952]: INFO nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Terminating instance [ 829.052079] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.148658] env[62952]: DEBUG nova.network.neutron [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.288388] env[62952]: DEBUG nova.network.neutron [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.303593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.390881] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.391480] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.394518] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.153s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.790904] env[62952]: DEBUG oslo_concurrency.lockutils [req-5acc88ac-d033-44c3-95b0-1180b3f7a980 req-e44aecc1-7a9e-404f-b7c4-6714082c0ae9 service nova] Releasing lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.791369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.791608] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.905791] env[62952]: DEBUG nova.compute.utils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.909569] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.909569] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.978952] env[62952]: DEBUG nova.policy [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3b70a4d7afa4caebd57270f89bf5ab0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fb2b7ad8e20454d83f9808b38ea26f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.242840] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367f127c-be97-4e76-9167-b4193d507513 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.250984] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e34536-0299-49f5-b350-af463964b5bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.281807] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882283cb-8336-49b9-bdf5-cd5766afff42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.291214] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245a6b7b-3cb9-496c-8a4d-556bc2b40f99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.310926] env[62952]: DEBUG nova.compute.provider_tree [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.333337] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.416508] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.544018] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.589976] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Successfully created port: 2877b7fb-e173-4285-9b70-fe63ad1c32dd {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.653796] env[62952]: DEBUG nova.compute.manager [req-ad078010-c0e8-4947-9778-139671219735 req-07fd8eab-442e-4c55-b798-b7f121c98b85 service nova] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Received event network-vif-deleted-4e8bbee8-553c-404c-a2a3-eaefff21a6f9 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.814160] env[62952]: DEBUG nova.scheduler.client.report [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.046743] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.047175] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.047369] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.047760] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f5ce80a-89dd-4dce-bb92-02ac11ee3c35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.056806] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6ccca1-21a7-4c92-9234-72fa04f7c7f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.078342] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4be2f0a4-5f57-4a80-9450-6d61d344799e could not be found. [ 831.078790] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.079057] env[62952]: INFO nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 831.079393] env[62952]: DEBUG oslo.service.loopingcall [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.079586] env[62952]: DEBUG nova.compute.manager [-] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.079769] env[62952]: DEBUG nova.network.neutron [-] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 831.104209] env[62952]: DEBUG nova.network.neutron [-] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.321240] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.924s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.321240] env[62952]: ERROR nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] Traceback (most recent call last): [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self.driver.spawn(context, instance, image_meta, [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self._vmops.spawn(context, instance, image_meta, injected_files, [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 831.321240] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] vm_ref = self.build_virtual_machine(instance, [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] vif_infos = vmwarevif.get_vif_info(self._session, [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] for vif in network_info: [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return self._sync_wrapper(fn, *args, **kwargs) [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self.wait() [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self[:] = self._gt.wait() [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return self._exit_event.wait() [ 831.321539] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] result = hub.switch() [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return self.greenlet.switch() [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] result = function(*args, **kwargs) [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] return func(*args, **kwargs) [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] raise e [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] nwinfo = self.network_api.allocate_for_instance( [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.321913] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] created_port_ids = self._update_ports_for_instance( [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] with excutils.save_and_reraise_exception(): [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] self.force_reraise() [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] raise self.value [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] updated_port = self._update_port( [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] _ensure_no_port_binding_failure(port) [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.322291] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] raise exception.PortBindingFailed(port_id=port['id']) [ 831.322775] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] nova.exception.PortBindingFailed: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. [ 831.322775] env[62952]: ERROR nova.compute.manager [instance: 2222309b-ce4e-46eb-a620-34553f644947] [ 831.322775] env[62952]: DEBUG nova.compute.utils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 831.323454] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Build of instance 2222309b-ce4e-46eb-a620-34553f644947 was re-scheduled: Binding failed for port bb5a6fe2-c4da-40ad-8df3-a44c9d125c05, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 831.323454] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 831.323813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquiring lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.324409] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Acquired lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.324528] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.326564] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.747s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.423674] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.450539] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.450793] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.450953] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.451147] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.451293] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.451442] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.451640] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.451797] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.451961] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.452250] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.452443] env[62952]: DEBUG nova.virt.hardware [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.453380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a8b532-63a4-4cbe-9065-a1f7d7d373ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.461571] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38907a23-aa28-4829-9c21-e3afc134e4ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.546206] env[62952]: ERROR nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 831.546206] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.546206] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.546206] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.546206] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.546206] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.546206] env[62952]: ERROR nova.compute.manager raise self.value [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.546206] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 831.546206] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.546206] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 831.546688] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.546688] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 831.546688] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 831.546688] env[62952]: ERROR nova.compute.manager [ 831.546688] env[62952]: Traceback (most recent call last): [ 831.546688] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 831.546688] env[62952]: listener.cb(fileno) [ 831.546688] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 831.546688] env[62952]: result = function(*args, **kwargs) [ 831.546688] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 831.546688] env[62952]: return func(*args, **kwargs) [ 831.546688] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 831.546688] env[62952]: raise e [ 831.546688] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.546688] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 831.546688] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.546688] env[62952]: created_port_ids = self._update_ports_for_instance( [ 831.546688] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.546688] env[62952]: with excutils.save_and_reraise_exception(): [ 831.546688] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.546688] env[62952]: self.force_reraise() [ 831.546688] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.546688] env[62952]: raise self.value [ 831.546688] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.546688] env[62952]: updated_port = self._update_port( [ 831.546688] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.546688] env[62952]: _ensure_no_port_binding_failure(port) [ 831.546688] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.546688] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 831.547519] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 831.547519] env[62952]: Removing descriptor: 19 [ 831.547519] env[62952]: ERROR nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Traceback (most recent call last): [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] yield resources [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self.driver.spawn(context, instance, image_meta, [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self._vmops.spawn(context, instance, image_meta, injected_files, [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 831.547519] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] vm_ref = self.build_virtual_machine(instance, [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] vif_infos = vmwarevif.get_vif_info(self._session, [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] for vif in network_info: [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return self._sync_wrapper(fn, *args, **kwargs) [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self.wait() [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self[:] = self._gt.wait() [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return self._exit_event.wait() [ 831.547863] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] result = hub.switch() [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return self.greenlet.switch() [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] result = function(*args, **kwargs) [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return func(*args, **kwargs) [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] raise e [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] nwinfo = self.network_api.allocate_for_instance( [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.548248] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] created_port_ids = self._update_ports_for_instance( [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] with excutils.save_and_reraise_exception(): [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self.force_reraise() [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] raise self.value [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] updated_port = self._update_port( [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] _ensure_no_port_binding_failure(port) [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.548630] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] raise exception.PortBindingFailed(port_id=port['id']) [ 831.548966] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 831.548966] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] [ 831.548966] env[62952]: INFO nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Terminating instance [ 831.549918] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquiring lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.549918] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquired lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.550102] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.606843] env[62952]: DEBUG nova.network.neutron [-] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.842300] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.926701] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.072291] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.109242] env[62952]: INFO nova.compute.manager [-] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Took 1.03 seconds to deallocate network for instance. [ 832.111425] env[62952]: DEBUG nova.compute.claims [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 832.111597] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.136903] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a0a74a-6385-4c61-953b-e628511ebba4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.144146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db20596-c3e8-439d-8b0e-bf31cb1a922d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.173584] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.175146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4c407d-a29b-4c88-bae2-466220e1823b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.182385] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00a7ea7-1501-416e-8194-229fae9963f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.195957] env[62952]: DEBUG nova.compute.provider_tree [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.431479] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Releasing lock "refresh_cache-2222309b-ce4e-46eb-a620-34553f644947" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.431758] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 832.433345] env[62952]: DEBUG nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.433345] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.455809] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.679473] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Releasing lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.679914] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.680122] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.680428] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc1ed144-4ed0-431f-988d-5c7f7cc0bcf5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.685716] env[62952]: DEBUG nova.compute.manager [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Received event network-changed-2877b7fb-e173-4285-9b70-fe63ad1c32dd {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.685894] env[62952]: DEBUG nova.compute.manager [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Refreshing instance network info cache due to event network-changed-2877b7fb-e173-4285-9b70-fe63ad1c32dd. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.686115] env[62952]: DEBUG oslo_concurrency.lockutils [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] Acquiring lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.686257] env[62952]: DEBUG oslo_concurrency.lockutils [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] Acquired lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.686414] env[62952]: DEBUG nova.network.neutron [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Refreshing network info cache for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.694065] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6cfe85-36a2-419e-a6e8-58438f5986c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.706479] env[62952]: DEBUG nova.scheduler.client.report [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.721599] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6f085d0c-c440-4bba-9bf8-cf2f70742748 could not be found. [ 832.721800] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.721977] env[62952]: INFO nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Took 0.04 seconds to destroy the instance on the hypervisor. [ 832.722227] env[62952]: DEBUG oslo.service.loopingcall [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.722979] env[62952]: DEBUG nova.compute.manager [-] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.723093] env[62952]: DEBUG nova.network.neutron [-] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.739384] env[62952]: DEBUG nova.network.neutron [-] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.958246] env[62952]: DEBUG nova.network.neutron [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.204039] env[62952]: DEBUG nova.network.neutron [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.210951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.211407] env[62952]: ERROR nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Traceback (most recent call last): [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self.driver.spawn(context, instance, image_meta, [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self._vmops.spawn(context, instance, image_meta, injected_files, [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] vm_ref = self.build_virtual_machine(instance, [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] vif_infos = vmwarevif.get_vif_info(self._session, [ 833.211407] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] for vif in network_info: [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] return self._sync_wrapper(fn, *args, **kwargs) [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self.wait() [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self[:] = self._gt.wait() [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] return self._exit_event.wait() [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] current.throw(*self._exc) [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.211788] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] result = function(*args, **kwargs) [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] return func(*args, **kwargs) [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] raise e [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] nwinfo = self.network_api.allocate_for_instance( [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] created_port_ids = self._update_ports_for_instance( [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] with excutils.save_and_reraise_exception(): [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] self.force_reraise() [ 833.212165] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] raise self.value [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] updated_port = self._update_port( [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] _ensure_no_port_binding_failure(port) [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] raise exception.PortBindingFailed(port_id=port['id']) [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] nova.exception.PortBindingFailed: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. [ 833.212681] env[62952]: ERROR nova.compute.manager [instance: 74361d48-0a82-453a-a4a1-45173c49e901] [ 833.212681] env[62952]: DEBUG nova.compute.utils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 833.213834] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Build of instance 74361d48-0a82-453a-a4a1-45173c49e901 was re-scheduled: Binding failed for port 3de940fa-7aef-49db-8fd2-f34b6e20e28f, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 833.214261] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 833.214490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquiring lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.214638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Acquired lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.214799] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.215830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.638s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.216018] env[62952]: DEBUG nova.objects.instance [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 833.240938] env[62952]: DEBUG nova.network.neutron [-] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.289316] env[62952]: DEBUG nova.network.neutron [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.460979] env[62952]: INFO nova.compute.manager [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] [instance: 2222309b-ce4e-46eb-a620-34553f644947] Took 1.03 seconds to deallocate network for instance. [ 833.737749] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.743203] env[62952]: INFO nova.compute.manager [-] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Took 1.02 seconds to deallocate network for instance. [ 833.745164] env[62952]: DEBUG nova.compute.claims [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 833.745392] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.792191] env[62952]: DEBUG oslo_concurrency.lockutils [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] Releasing lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.792348] env[62952]: DEBUG nova.compute.manager [req-dfe47fe2-1937-4f66-bf12-a0c69309fbf8 req-72801af3-2d07-42ab-9d56-fafb59224cda service nova] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Received event network-vif-deleted-2877b7fb-e173-4285-9b70-fe63ad1c32dd {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.801472] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.226692] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8025c159-adf9-4374-954f-3ce10d22188f tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.227901] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.020s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.304151] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Releasing lock "refresh_cache-74361d48-0a82-453a-a4a1-45173c49e901" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.304390] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 834.304577] env[62952]: DEBUG nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.304738] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.320690] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.498355] env[62952]: INFO nova.scheduler.client.report [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Deleted allocations for instance 2222309b-ce4e-46eb-a620-34553f644947 [ 834.825804] env[62952]: DEBUG nova.network.neutron [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.973256] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3d3fdb-ed59-49c5-a907-0e46a153bc1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.981215] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa96f1e-9f54-482d-ba58-86d8854163b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.009229] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3395ce28-ebd0-48a6-99e1-930406bd6890 tempest-ServersTestManualDisk-1814691432 tempest-ServersTestManualDisk-1814691432-project-member] Lock "2222309b-ce4e-46eb-a620-34553f644947" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 164.004s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.010860] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85e6c73-54be-44db-8a81-a306e334f92a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.018872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1316d3e-47ae-4b46-9f45-40cc84aad473 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.032139] env[62952]: DEBUG nova.compute.provider_tree [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.329052] env[62952]: INFO nova.compute.manager [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] [instance: 74361d48-0a82-453a-a4a1-45173c49e901] Took 1.02 seconds to deallocate network for instance. [ 835.513965] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.535063] env[62952]: DEBUG nova.scheduler.client.report [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.038025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.040908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.813s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.041579] env[62952]: ERROR nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Traceback (most recent call last): [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self.driver.spawn(context, instance, image_meta, [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] vm_ref = self.build_virtual_machine(instance, [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] vif_infos = vmwarevif.get_vif_info(self._session, [ 836.041579] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] for vif in network_info: [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return self._sync_wrapper(fn, *args, **kwargs) [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self.wait() [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self[:] = self._gt.wait() [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return self._exit_event.wait() [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] result = hub.switch() [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 836.041985] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return self.greenlet.switch() [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] result = function(*args, **kwargs) [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] return func(*args, **kwargs) [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] raise e [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] nwinfo = self.network_api.allocate_for_instance( [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] created_port_ids = self._update_ports_for_instance( [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] with excutils.save_and_reraise_exception(): [ 836.042426] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] self.force_reraise() [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] raise self.value [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] updated_port = self._update_port( [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] _ensure_no_port_binding_failure(port) [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] raise exception.PortBindingFailed(port_id=port['id']) [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] nova.exception.PortBindingFailed: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. [ 836.042820] env[62952]: ERROR nova.compute.manager [instance: 91660994-feca-40d8-8b75-d3c79110a19e] [ 836.043193] env[62952]: DEBUG nova.compute.utils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 836.043521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.449s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.045496] env[62952]: INFO nova.compute.claims [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.048323] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Build of instance 91660994-feca-40d8-8b75-d3c79110a19e was re-scheduled: Binding failed for port d78daeab-1f1d-441c-8b47-dea2fbe31ad4, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 836.048812] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 836.049078] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquiring lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.049255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Acquired lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.049439] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.360313] env[62952]: INFO nova.scheduler.client.report [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Deleted allocations for instance 74361d48-0a82-453a-a4a1-45173c49e901 [ 836.574522] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.683541] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.869430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-332a3e1c-4c5e-4a35-bf18-965aec531699 tempest-ImagesOneServerTestJSON-74964259 tempest-ImagesOneServerTestJSON-74964259-project-member] Lock "74361d48-0a82-453a-a4a1-45173c49e901" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.345s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.187741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Releasing lock "refresh_cache-91660994-feca-40d8-8b75-d3c79110a19e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.187741] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 837.187741] env[62952]: DEBUG nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 837.187741] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 837.207407] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.375022] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.402223] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef45db1-927c-41e2-a844-6b2b0ad8c27d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.410477] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f554b4-c26e-4701-91fa-f0cb34a5f3b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.447044] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c473ce-8036-4ea6-b838-3f9b6758a169 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.459483] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921b75af-8ba6-4a2c-baaa-165273f897e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.473312] env[62952]: DEBUG nova.compute.provider_tree [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.709577] env[62952]: DEBUG nova.network.neutron [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.903915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.976162] env[62952]: DEBUG nova.scheduler.client.report [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.217115] env[62952]: INFO nova.compute.manager [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] [instance: 91660994-feca-40d8-8b75-d3c79110a19e] Took 1.03 seconds to deallocate network for instance. [ 838.481985] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.482789] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.485534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.895s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.485802] env[62952]: DEBUG nova.objects.instance [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lazy-loading 'resources' on Instance uuid 9567865d-41ae-4379-8313-e87bfc92fba8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.989676] env[62952]: DEBUG nova.compute.utils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.991705] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.991883] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.072154] env[62952]: DEBUG nova.policy [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '592ab470a2504c459a9f20f6e3c7a3bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71cf6551882a4c34ba621ee9b76efb8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.249542] env[62952]: INFO nova.scheduler.client.report [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Deleted allocations for instance 91660994-feca-40d8-8b75-d3c79110a19e [ 839.317583] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748048dd-d607-4b94-b2fe-5a91c0fbb35c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.326764] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16277fdb-8c6a-4073-8178-6c53d4e4ad5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.362889] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3edcb2e-48ba-4c90-96cf-1e2bc75eac46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.371223] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e860da71-89c6-47a2-891d-af49198b5c1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.388105] env[62952]: DEBUG nova.compute.provider_tree [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.493532] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.719283] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Successfully created port: af77a34b-4d47-42de-97f3-b3afc5627973 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.768629] env[62952]: DEBUG oslo_concurrency.lockutils [None req-03919a2b-2872-4784-bb8f-6ea9a428a0fa tempest-ServerRescueTestJSON-230026192 tempest-ServerRescueTestJSON-230026192-project-member] Lock "91660994-feca-40d8-8b75-d3c79110a19e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.380s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.897106] env[62952]: DEBUG nova.scheduler.client.report [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.273760] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.401874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.915s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.403430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.792s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.405789] env[62952]: INFO nova.compute.claims [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.429391] env[62952]: INFO nova.scheduler.client.report [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Deleted allocations for instance 9567865d-41ae-4379-8313-e87bfc92fba8 [ 840.508553] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.541358] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.541610] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.541764] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.541940] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.542102] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.542654] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.542654] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.542654] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.542802] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.542901] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.543082] env[62952]: DEBUG nova.virt.hardware [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.543938] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df99132b-6e27-4e8f-b676-abe206858cb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.553290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b6de5a-a430-4ace-a6f7-1d9c95a991e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.798634] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.938410] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b6825f3e-021a-400a-9b7b-e410395fa6c9 tempest-ServerShowV254Test-1622428837 tempest-ServerShowV254Test-1622428837-project-member] Lock "9567865d-41ae-4379-8313-e87bfc92fba8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.094s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.727102] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95799d3e-b84a-4271-99f7-6eba10701c4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.734100] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9387fe87-6439-4fb7-b43d-33d540859e76 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.772370] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b87f2e2-914f-4372-9310-17173f33df4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.779471] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf0fcd0-9f65-4a7b-b7f1-c364602cb7cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.792421] env[62952]: DEBUG nova.compute.provider_tree [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.296079] env[62952]: DEBUG nova.scheduler.client.report [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.335524] env[62952]: DEBUG nova.compute.manager [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Received event network-changed-af77a34b-4d47-42de-97f3-b3afc5627973 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.335784] env[62952]: DEBUG nova.compute.manager [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Refreshing instance network info cache due to event network-changed-af77a34b-4d47-42de-97f3-b3afc5627973. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.336087] env[62952]: DEBUG oslo_concurrency.lockutils [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] Acquiring lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.336300] env[62952]: DEBUG oslo_concurrency.lockutils [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] Acquired lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.336516] env[62952]: DEBUG nova.network.neutron [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Refreshing network info cache for port af77a34b-4d47-42de-97f3-b3afc5627973 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.681454] env[62952]: ERROR nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 842.681454] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.681454] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.681454] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.681454] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.681454] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.681454] env[62952]: ERROR nova.compute.manager raise self.value [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.681454] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 842.681454] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.681454] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 842.681931] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.681931] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 842.681931] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 842.681931] env[62952]: ERROR nova.compute.manager [ 842.681931] env[62952]: Traceback (most recent call last): [ 842.681931] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 842.681931] env[62952]: listener.cb(fileno) [ 842.681931] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 842.681931] env[62952]: result = function(*args, **kwargs) [ 842.681931] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 842.681931] env[62952]: return func(*args, **kwargs) [ 842.681931] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 842.681931] env[62952]: raise e [ 842.681931] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.681931] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 842.681931] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.681931] env[62952]: created_port_ids = self._update_ports_for_instance( [ 842.681931] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.681931] env[62952]: with excutils.save_and_reraise_exception(): [ 842.681931] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.681931] env[62952]: self.force_reraise() [ 842.681931] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.681931] env[62952]: raise self.value [ 842.682576] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.682576] env[62952]: updated_port = self._update_port( [ 842.682576] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.682576] env[62952]: _ensure_no_port_binding_failure(port) [ 842.682576] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.682576] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 842.682576] env[62952]: nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 842.682576] env[62952]: Removing descriptor: 19 [ 842.682807] env[62952]: ERROR nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Traceback (most recent call last): [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] yield resources [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self.driver.spawn(context, instance, image_meta, [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] vm_ref = self.build_virtual_machine(instance, [ 842.682807] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] vif_infos = vmwarevif.get_vif_info(self._session, [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] for vif in network_info: [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return self._sync_wrapper(fn, *args, **kwargs) [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self.wait() [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self[:] = self._gt.wait() [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return self._exit_event.wait() [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 842.683216] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] result = hub.switch() [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return self.greenlet.switch() [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] result = function(*args, **kwargs) [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return func(*args, **kwargs) [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] raise e [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] nwinfo = self.network_api.allocate_for_instance( [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] created_port_ids = self._update_ports_for_instance( [ 842.683584] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] with excutils.save_and_reraise_exception(): [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self.force_reraise() [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] raise self.value [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] updated_port = self._update_port( [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] _ensure_no_port_binding_failure(port) [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] raise exception.PortBindingFailed(port_id=port['id']) [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 842.683961] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] [ 842.684356] env[62952]: INFO nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Terminating instance [ 842.688778] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.803129] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.804143] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.806928] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.429s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.860847] env[62952]: DEBUG nova.network.neutron [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.958985] env[62952]: DEBUG nova.network.neutron [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.314564] env[62952]: DEBUG nova.compute.utils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.315991] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.316204] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.408125] env[62952]: DEBUG nova.policy [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '969a00a745194f36adea881ba34cc9c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8b59dc3b814d6ca5fa1d03d04660c6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.462322] env[62952]: DEBUG oslo_concurrency.lockutils [req-d913b763-475b-427e-b2c4-59d10e4b005c req-f23f6efa-d07d-42ce-b0c8-f70b0f17b451 service nova] Releasing lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.462701] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquired lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.462798] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.699336] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d80e31b-d107-4ce9-815c-5e65ac2dc9c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.707272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567ba983-a74f-49c0-a8d4-b71e77cca044 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.752294] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1131e5c8-34b8-44a7-9ef9-862a98cd37a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.760122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6dbc21-6e80-4e20-95f1-2b72f56fed1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.775023] env[62952]: DEBUG nova.compute.provider_tree [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.819972] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.914172] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Successfully created port: d90be0de-44f3-429b-831a-f29cb3a5182a {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.010130] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.100404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "34f5995f-692c-4f0d-8b15-0d388df7e34e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.100514] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.191794] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.278942] env[62952]: DEBUG nova.scheduler.client.report [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.372019] env[62952]: DEBUG nova.compute.manager [req-c4e0d25e-586e-46c0-8595-5de8de28d791 req-6db5ee37-e1a7-42af-a454-a3a236ab80f9 service nova] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Received event network-vif-deleted-af77a34b-4d47-42de-97f3-b3afc5627973 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.693995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Releasing lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.694480] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.694744] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.695076] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b68951ab-6dde-4144-a20b-51b1a870691c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.708121] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9c88e1-30d1-43d6-a2bc-d5f6da1c554f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.738249] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95c07667-e9d8-4cc3-b135-f80d196d179f could not be found. [ 844.738534] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.738734] env[62952]: INFO nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 844.739073] env[62952]: DEBUG oslo.service.loopingcall [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.739345] env[62952]: DEBUG nova.compute.manager [-] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.739452] env[62952]: DEBUG nova.network.neutron [-] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.765910] env[62952]: DEBUG nova.network.neutron [-] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.787677] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.788207] env[62952]: ERROR nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Traceback (most recent call last): [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self.driver.spawn(context, instance, image_meta, [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self._vmops.spawn(context, instance, image_meta, injected_files, [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] vm_ref = self.build_virtual_machine(instance, [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] vif_infos = vmwarevif.get_vif_info(self._session, [ 844.788207] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] for vif in network_info: [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return self._sync_wrapper(fn, *args, **kwargs) [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self.wait() [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self[:] = self._gt.wait() [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return self._exit_event.wait() [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] result = hub.switch() [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 844.788764] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return self.greenlet.switch() [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] result = function(*args, **kwargs) [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] return func(*args, **kwargs) [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] raise e [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] nwinfo = self.network_api.allocate_for_instance( [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] created_port_ids = self._update_ports_for_instance( [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] with excutils.save_and_reraise_exception(): [ 844.789170] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] self.force_reraise() [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] raise self.value [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] updated_port = self._update_port( [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] _ensure_no_port_binding_failure(port) [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] raise exception.PortBindingFailed(port_id=port['id']) [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] nova.exception.PortBindingFailed: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. [ 844.789610] env[62952]: ERROR nova.compute.manager [instance: c7771b47-cbe8-4696-9a21-9204501c8877] [ 844.789952] env[62952]: DEBUG nova.compute.utils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 844.790624] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.587s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.790827] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.790981] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 844.791280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.458s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.792950] env[62952]: INFO nova.compute.claims [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.795638] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Build of instance c7771b47-cbe8-4696-9a21-9204501c8877 was re-scheduled: Binding failed for port fa638cdd-dbbc-466f-9d99-c3a874f1ca12, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 844.796078] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 844.796320] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquiring lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.800028] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Acquired lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.800028] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.800028] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf94e846-1a61-4ed1-8f90-f39e6c8b7287 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.807239] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f602b728-1e2e-4655-8a80-feb0df803cc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.826304] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d3940c-bf11-480e-ba6d-360fbac6b964 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.831683] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.836042] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117e67be-478c-4d97-8fa0-96e17962a4f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.871012] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181485MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 844.871187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.879906] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.880151] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.880306] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.880487] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.880630] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.880775] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.880976] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.881143] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.881415] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.881480] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.881622] env[62952]: DEBUG nova.virt.hardware [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.882667] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be0e9b8-861a-4d7a-8960-a66cdd5b89b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.890950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce40116-408b-4645-b0e2-01cf14f0d5f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.269231] env[62952]: DEBUG nova.network.neutron [-] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.340390] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.556206] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.773321] env[62952]: INFO nova.compute.manager [-] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Took 1.03 seconds to deallocate network for instance. [ 845.774369] env[62952]: DEBUG nova.compute.claims [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 845.774699] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.016270] env[62952]: ERROR nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 846.016270] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 846.016270] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 846.016270] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 846.016270] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 846.016270] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 846.016270] env[62952]: ERROR nova.compute.manager raise self.value [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 846.016270] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 846.016270] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 846.016270] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 846.017012] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 846.017012] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 846.017012] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 846.017012] env[62952]: ERROR nova.compute.manager [ 846.017012] env[62952]: Traceback (most recent call last): [ 846.017012] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 846.017012] env[62952]: listener.cb(fileno) [ 846.017012] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 846.017012] env[62952]: result = function(*args, **kwargs) [ 846.017012] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 846.017012] env[62952]: return func(*args, **kwargs) [ 846.017012] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 846.017012] env[62952]: raise e [ 846.017012] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 846.017012] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 846.017012] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 846.017012] env[62952]: created_port_ids = self._update_ports_for_instance( [ 846.017012] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 846.017012] env[62952]: with excutils.save_and_reraise_exception(): [ 846.017012] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 846.017012] env[62952]: self.force_reraise() [ 846.017012] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 846.017012] env[62952]: raise self.value [ 846.017012] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 846.017012] env[62952]: updated_port = self._update_port( [ 846.017012] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 846.017012] env[62952]: _ensure_no_port_binding_failure(port) [ 846.017012] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 846.017012] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 846.017897] env[62952]: nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 846.017897] env[62952]: Removing descriptor: 19 [ 846.017897] env[62952]: ERROR nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Traceback (most recent call last): [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] yield resources [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self.driver.spawn(context, instance, image_meta, [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 846.017897] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] vm_ref = self.build_virtual_machine(instance, [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] vif_infos = vmwarevif.get_vif_info(self._session, [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] for vif in network_info: [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return self._sync_wrapper(fn, *args, **kwargs) [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self.wait() [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self[:] = self._gt.wait() [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return self._exit_event.wait() [ 846.018358] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] result = hub.switch() [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return self.greenlet.switch() [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] result = function(*args, **kwargs) [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return func(*args, **kwargs) [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] raise e [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] nwinfo = self.network_api.allocate_for_instance( [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 846.018787] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] created_port_ids = self._update_ports_for_instance( [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] with excutils.save_and_reraise_exception(): [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self.force_reraise() [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] raise self.value [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] updated_port = self._update_port( [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] _ensure_no_port_binding_failure(port) [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 846.019793] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] raise exception.PortBindingFailed(port_id=port['id']) [ 846.021016] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 846.021016] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] [ 846.021016] env[62952]: INFO nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Terminating instance [ 846.021016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquiring lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.021016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquired lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.021016] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.106929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Releasing lock "refresh_cache-c7771b47-cbe8-4696-9a21-9204501c8877" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.106929] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 846.106929] env[62952]: DEBUG nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.106929] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.106929] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.168021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ec98eb-96c6-4796-b963-0a3d964101cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.177321] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9254cae3-36c2-4671-87c0-b53949979919 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.206341] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0455238d-77e3-4fb8-9dc2-98eaf0fa13b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.213485] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef339b4-dc6c-4ad6-b788-b0c51e5241f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.226789] env[62952]: DEBUG nova.compute.provider_tree [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.401056] env[62952]: DEBUG nova.compute.manager [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Received event network-changed-d90be0de-44f3-429b-831a-f29cb3a5182a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.401252] env[62952]: DEBUG nova.compute.manager [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Refreshing instance network info cache due to event network-changed-d90be0de-44f3-429b-831a-f29cb3a5182a. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.401434] env[62952]: DEBUG oslo_concurrency.lockutils [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] Acquiring lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.541727] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.590478] env[62952]: DEBUG nova.network.neutron [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.616595] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.731993] env[62952]: DEBUG nova.scheduler.client.report [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.092857] env[62952]: INFO nova.compute.manager [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] [instance: c7771b47-cbe8-4696-9a21-9204501c8877] Took 1.03 seconds to deallocate network for instance. [ 847.119562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Releasing lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.119562] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.119699] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 847.120125] env[62952]: DEBUG oslo_concurrency.lockutils [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] Acquired lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.120636] env[62952]: DEBUG nova.network.neutron [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Refreshing network info cache for port d90be0de-44f3-429b-831a-f29cb3a5182a {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.121862] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bb56b19-8a27-4b95-8623-448c08d51764 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.133136] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c950fcf-284e-42f0-adf7-e3f50169f3eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.158156] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca7fc13d-e9be-481c-ae58-78377627800c could not be found. [ 847.158442] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.158893] env[62952]: INFO nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 847.158956] env[62952]: DEBUG oslo.service.loopingcall [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.159192] env[62952]: DEBUG nova.compute.manager [-] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.159308] env[62952]: DEBUG nova.network.neutron [-] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.175864] env[62952]: DEBUG nova.network.neutron [-] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.235165] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.235689] env[62952]: DEBUG nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.238549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.935s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.240160] env[62952]: INFO nova.compute.claims [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.648237] env[62952]: DEBUG nova.network.neutron [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.678979] env[62952]: DEBUG nova.network.neutron [-] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.703166] env[62952]: DEBUG nova.network.neutron [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.745663] env[62952]: DEBUG nova.compute.utils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.751946] env[62952]: DEBUG nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 848.140156] env[62952]: INFO nova.scheduler.client.report [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Deleted allocations for instance c7771b47-cbe8-4696-9a21-9204501c8877 [ 848.181986] env[62952]: INFO nova.compute.manager [-] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Took 1.02 seconds to deallocate network for instance. [ 848.185118] env[62952]: DEBUG nova.compute.claims [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 848.185300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.206593] env[62952]: DEBUG oslo_concurrency.lockutils [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] Releasing lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.206593] env[62952]: DEBUG nova.compute.manager [req-16571bb0-74ca-4b71-b7c4-f3b5526f269f req-91d9f1d7-3adb-4af7-8e0e-57b76ca61506 service nova] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Received event network-vif-deleted-d90be0de-44f3-429b-831a-f29cb3a5182a {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.256018] env[62952]: DEBUG nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 848.650667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0dc9e9ab-71b0-41c1-b9c7-4caf8116d044 tempest-ImagesNegativeTestJSON-83741465 tempest-ImagesNegativeTestJSON-83741465-project-member] Lock "c7771b47-cbe8-4696-9a21-9204501c8877" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 171.936s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.661802] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7267e02-6035-426a-b8de-ae7c25ce1554 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.671299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e14a55-ef0f-4e61-b2f6-87964e06d792 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.710434] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6c4917-f69a-4992-b393-dec6ac2f21b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.717881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2112924e-57df-4dd7-9164-1e1461ad14b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.731072] env[62952]: DEBUG nova.compute.provider_tree [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.157567] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.236015] env[62952]: DEBUG nova.scheduler.client.report [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.265459] env[62952]: DEBUG nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.291235] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.291461] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.291615] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.291897] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.292064] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.292214] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.292418] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.292572] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.292739] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.292901] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.293071] env[62952]: DEBUG nova.virt.hardware [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.294447] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43cf430-e187-46c2-b1bf-310b28d9f3c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.302775] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1dc3bac-9308-4ec7-84be-a3a369bd2f9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.316625] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.322218] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Creating folder: Project (48d1df38ff244d87a9e97cee7025c3dd). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.322811] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ac06f1d-d189-4638-9607-fc6519605d89 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.336587] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Created folder: Project (48d1df38ff244d87a9e97cee7025c3dd) in parent group-v290852. [ 849.336587] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Creating folder: Instances. Parent ref: group-v290870. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.336587] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e66bbeb3-5576-4eec-b3ae-c19a394da374 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.349015] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Created folder: Instances in parent group-v290870. [ 849.349015] env[62952]: DEBUG oslo.service.loopingcall [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.349015] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.349015] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25d3c013-8f0e-4031-90d1-b41881a9862c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.364633] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.364633] env[62952]: value = "task-1367097" [ 849.364633] env[62952]: _type = "Task" [ 849.364633] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.372236] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367097, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.691084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.743099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.743099] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.747225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.634s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.878566] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367097, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.130875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "4983f354-77dc-4e84-969b-4c74441fa568" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.130875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "4983f354-77dc-4e84-969b-4c74441fa568" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.257923] env[62952]: DEBUG nova.compute.utils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.260367] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.260573] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.377629] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367097, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.493166] env[62952]: DEBUG nova.policy [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88ff36903a64472aac028f0367fa59f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9bd4f56d8d6415c9f5d98fe6d2562ed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.560983] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f5d0e2-f05b-4865-81ff-6875a519e679 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.569243] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f73b5f-1eaa-4bd3-a8c9-6c34f5e9737d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.601921] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f5fbd5-755c-46c0-b903-c97f97b8e0f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.609921] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2423f3b-f4e2-41ec-9d26-3a0cc9991ca7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.623189] env[62952]: DEBUG nova.compute.provider_tree [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.765288] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.882043] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367097, 'name': CreateVM_Task, 'duration_secs': 1.317266} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.882196] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.882698] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.883391] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.883771] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.885082] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abc116da-418c-442c-b089-9e553ec22ad4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.894099] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 850.894099] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52143eed-e9ea-423b-15d2-84a02c7d222f" [ 850.894099] env[62952]: _type = "Task" [ 850.894099] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.900353] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52143eed-e9ea-423b-15d2-84a02c7d222f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.967827] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Successfully created port: a70da931-6354-49a6-aa2c-7d085de6ae67 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.129221] env[62952]: DEBUG nova.scheduler.client.report [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.274335] env[62952]: INFO nova.virt.block_device [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Booting with volume 768ba74c-8c9a-45ea-b9a6-343ae3ebf625 at /dev/sda [ 851.328019] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7276068-f395-4084-9917-266fe255568d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.338086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a25e57-d477-4fea-99b2-8cff95bb6332 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.361647] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d23a2562-f60e-4e2b-8121-ed4926f3b990 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.369582] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2bf304d-630f-4223-a66e-315280d11f3e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.392276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b090f6-d5ed-4303-bb0a-4bac0f3d0f1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.404082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457e73a6-a1da-43c5-8760-8870823a8dd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.407130] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52143eed-e9ea-423b-15d2-84a02c7d222f, 'name': SearchDatastore_Task, 'duration_secs': 0.010181} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.407408] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.407633] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.407859] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.407999] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.408185] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.408774] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48361051-a6ed-46bc-8d97-d5e76c5f1912 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.419069] env[62952]: DEBUG nova.virt.block_device [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Updating existing volume attachment record: 6dc96f20-7855-45a9-ab9c-3abe98a258d3 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 851.423802] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.423802] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.424376] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c1881a1-163d-43ba-874c-92fdc9c2d013 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.436382] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 851.436382] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]525ae001-5285-0348-6175-3ac5801e2ef4" [ 851.436382] env[62952]: _type = "Task" [ 851.436382] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.443615] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525ae001-5285-0348-6175-3ac5801e2ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.633252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.634068] env[62952]: ERROR nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Traceback (most recent call last): [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self.driver.spawn(context, instance, image_meta, [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] vm_ref = self.build_virtual_machine(instance, [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] vif_infos = vmwarevif.get_vif_info(self._session, [ 851.634068] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] for vif in network_info: [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] return self._sync_wrapper(fn, *args, **kwargs) [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self.wait() [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self[:] = self._gt.wait() [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] return self._exit_event.wait() [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] current.throw(*self._exc) [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 851.634346] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] result = function(*args, **kwargs) [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] return func(*args, **kwargs) [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] raise e [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] nwinfo = self.network_api.allocate_for_instance( [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] created_port_ids = self._update_ports_for_instance( [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] with excutils.save_and_reraise_exception(): [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] self.force_reraise() [ 851.634636] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] raise self.value [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] updated_port = self._update_port( [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] _ensure_no_port_binding_failure(port) [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] raise exception.PortBindingFailed(port_id=port['id']) [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] nova.exception.PortBindingFailed: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. [ 851.634894] env[62952]: ERROR nova.compute.manager [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] [ 851.634894] env[62952]: DEBUG nova.compute.utils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 851.636417] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.891s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.639424] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Build of instance 4be2f0a4-5f57-4a80-9450-6d61d344799e was re-scheduled: Binding failed for port 4e8bbee8-553c-404c-a2a3-eaefff21a6f9, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 851.640544] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 851.640544] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.640544] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.640544] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.857085] env[62952]: DEBUG nova.compute.manager [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Received event network-changed-a70da931-6354-49a6-aa2c-7d085de6ae67 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.857085] env[62952]: DEBUG nova.compute.manager [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Refreshing instance network info cache due to event network-changed-a70da931-6354-49a6-aa2c-7d085de6ae67. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.857085] env[62952]: DEBUG oslo_concurrency.lockutils [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] Acquiring lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.857085] env[62952]: DEBUG oslo_concurrency.lockutils [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] Acquired lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.857085] env[62952]: DEBUG nova.network.neutron [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Refreshing network info cache for port a70da931-6354-49a6-aa2c-7d085de6ae67 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.949023] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525ae001-5285-0348-6175-3ac5801e2ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.014158} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.949023] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8c1227a-264b-4690-b0fb-86679bc54d1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.953138] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 851.953138] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52237dbc-f0ac-90ed-006f-d17620b9152a" [ 851.953138] env[62952]: _type = "Task" [ 851.953138] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.959444] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52237dbc-f0ac-90ed-006f-d17620b9152a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.139914] env[62952]: ERROR nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 852.139914] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 852.139914] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 852.139914] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 852.139914] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 852.139914] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 852.139914] env[62952]: ERROR nova.compute.manager raise self.value [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 852.139914] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 852.139914] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 852.139914] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 852.140407] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 852.140407] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 852.140407] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 852.140407] env[62952]: ERROR nova.compute.manager [ 852.140407] env[62952]: Traceback (most recent call last): [ 852.140407] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 852.140407] env[62952]: listener.cb(fileno) [ 852.140407] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 852.140407] env[62952]: result = function(*args, **kwargs) [ 852.140407] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 852.140407] env[62952]: return func(*args, **kwargs) [ 852.140407] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 852.140407] env[62952]: raise e [ 852.140407] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 852.140407] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 852.140407] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 852.140407] env[62952]: created_port_ids = self._update_ports_for_instance( [ 852.140407] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 852.140407] env[62952]: with excutils.save_and_reraise_exception(): [ 852.140407] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 852.140407] env[62952]: self.force_reraise() [ 852.140407] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 852.140407] env[62952]: raise self.value [ 852.140407] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 852.140407] env[62952]: updated_port = self._update_port( [ 852.140407] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 852.140407] env[62952]: _ensure_no_port_binding_failure(port) [ 852.140407] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 852.140407] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 852.141236] env[62952]: nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 852.141236] env[62952]: Removing descriptor: 19 [ 852.159797] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.236269] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.375574] env[62952]: DEBUG nova.network.neutron [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.411405] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5e52e7-bb53-4e50-a1cf-4f830c5f310b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.421112] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc98efd-0bc7-4a80-9aaa-ab2e8212ea22 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.450343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5dc312-d31b-4366-a635-7802014eaf70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.453413] env[62952]: DEBUG nova.network.neutron [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.464747] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b9fc40-f5df-4d89-8190-43263c478a12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.468281] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52237dbc-f0ac-90ed-006f-d17620b9152a, 'name': SearchDatastore_Task, 'duration_secs': 0.009262} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.469114] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.469373] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 34b6c13c-c95b-4cc2-b6eb-cc9374535aec/34b6c13c-c95b-4cc2-b6eb-cc9374535aec.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.469960] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2600695c-4eda-4d99-9091-b1de3e2466c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.482424] env[62952]: DEBUG nova.compute.provider_tree [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.489988] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 852.489988] env[62952]: value = "task-1367098" [ 852.489988] env[62952]: _type = "Task" [ 852.489988] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.498248] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.739226] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "refresh_cache-4be2f0a4-5f57-4a80-9450-6d61d344799e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.739871] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 852.739871] env[62952]: DEBUG nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.740028] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.757041] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.958345] env[62952]: DEBUG oslo_concurrency.lockutils [req-277f715b-9b3b-4f8e-a27b-43b34e62554b req-ae1fc18b-d02e-4782-a61b-7f4dfc23fd56 service nova] Releasing lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.986574] env[62952]: DEBUG nova.scheduler.client.report [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.999813] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463449} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.000065] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 34b6c13c-c95b-4cc2-b6eb-cc9374535aec/34b6c13c-c95b-4cc2-b6eb-cc9374535aec.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.000274] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.001049] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1385840-cd82-4735-bd44-e23b009174bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.007700] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 853.007700] env[62952]: value = "task-1367099" [ 853.007700] env[62952]: _type = "Task" [ 853.007700] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.017343] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367099, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.260625] env[62952]: DEBUG nova.network.neutron [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.491215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.855s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.491867] env[62952]: ERROR nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Traceback (most recent call last): [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self.driver.spawn(context, instance, image_meta, [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self._vmops.spawn(context, instance, image_meta, injected_files, [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] vm_ref = self.build_virtual_machine(instance, [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] vif_infos = vmwarevif.get_vif_info(self._session, [ 853.491867] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] for vif in network_info: [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return self._sync_wrapper(fn, *args, **kwargs) [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self.wait() [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self[:] = self._gt.wait() [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return self._exit_event.wait() [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] result = hub.switch() [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 853.492169] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return self.greenlet.switch() [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] result = function(*args, **kwargs) [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] return func(*args, **kwargs) [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] raise e [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] nwinfo = self.network_api.allocate_for_instance( [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] created_port_ids = self._update_ports_for_instance( [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] with excutils.save_and_reraise_exception(): [ 853.492574] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] self.force_reraise() [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] raise self.value [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] updated_port = self._update_port( [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] _ensure_no_port_binding_failure(port) [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] raise exception.PortBindingFailed(port_id=port['id']) [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] nova.exception.PortBindingFailed: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. [ 853.492896] env[62952]: ERROR nova.compute.manager [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] [ 853.493210] env[62952]: DEBUG nova.compute.utils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 853.494185] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.456s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.495599] env[62952]: INFO nova.compute.claims [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.501020] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Build of instance 6f085d0c-c440-4bba-9bf8-cf2f70742748 was re-scheduled: Binding failed for port 2877b7fb-e173-4285-9b70-fe63ad1c32dd, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 853.501020] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 853.501020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquiring lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.501020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Acquired lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.501237] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.521145] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367099, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061579} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.521686] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.523135] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.523597] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.524089] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.524251] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.524430] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.524567] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.524702] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.524903] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.525106] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.525241] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.525395] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.525558] env[62952]: DEBUG nova.virt.hardware [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.527526] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607f8950-1788-4646-b5fc-c4dde25d3699 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.536723] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9b69e0-a26e-4a7e-9eb6-3dfa52c77ebd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.547013] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18a9598-ac96-4810-89ef-b50f66423d70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.559864] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 34b6c13c-c95b-4cc2-b6eb-cc9374535aec/34b6c13c-c95b-4cc2-b6eb-cc9374535aec.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.560733] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a480b103-00e6-4a12-b013-2c56c3473af0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.586400] env[62952]: ERROR nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Traceback (most recent call last): [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] yield resources [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self.driver.spawn(context, instance, image_meta, [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self._vmops.spawn(context, instance, image_meta, injected_files, [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] vm_ref = self.build_virtual_machine(instance, [ 853.586400] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] vif_infos = vmwarevif.get_vif_info(self._session, [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] for vif in network_info: [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] return self._sync_wrapper(fn, *args, **kwargs) [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self.wait() [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self[:] = self._gt.wait() [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] return self._exit_event.wait() [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 853.586802] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] current.throw(*self._exc) [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] result = function(*args, **kwargs) [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] return func(*args, **kwargs) [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] raise e [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] nwinfo = self.network_api.allocate_for_instance( [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] created_port_ids = self._update_ports_for_instance( [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] with excutils.save_and_reraise_exception(): [ 853.587183] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self.force_reraise() [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] raise self.value [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] updated_port = self._update_port( [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] _ensure_no_port_binding_failure(port) [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] raise exception.PortBindingFailed(port_id=port['id']) [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 853.587505] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] [ 853.587505] env[62952]: INFO nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Terminating instance [ 853.590091] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 853.590091] env[62952]: value = "task-1367100" [ 853.590091] env[62952]: _type = "Task" [ 853.590091] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.590599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquiring lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.590858] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquired lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.590912] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.600342] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367100, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.762665] env[62952]: INFO nova.compute.manager [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 4be2f0a4-5f57-4a80-9450-6d61d344799e] Took 1.02 seconds to deallocate network for instance. [ 853.882215] env[62952]: DEBUG nova.compute.manager [req-32e5990f-93ea-4e7b-b0f6-ea7e7afc3136 req-ffdefcd8-f324-4ced-918c-2113a848f5a4 service nova] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Received event network-vif-deleted-a70da931-6354-49a6-aa2c-7d085de6ae67 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.033751] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.106470] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367100, 'name': ReconfigVM_Task, 'duration_secs': 0.267375} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.106752] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 34b6c13c-c95b-4cc2-b6eb-cc9374535aec/34b6c13c-c95b-4cc2-b6eb-cc9374535aec.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.107474] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4d30149-4c94-443b-a502-ef2c4473d242 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.115262] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 854.115262] env[62952]: value = "task-1367101" [ 854.115262] env[62952]: _type = "Task" [ 854.115262] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.124525] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367101, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.125907] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.190931] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.247078] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.625167] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367101, 'name': Rename_Task, 'duration_secs': 0.147292} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.627563] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.628751] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2877de4e-118a-4d6b-91fa-400a075ce824 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.635786] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 854.635786] env[62952]: value = "task-1367102" [ 854.635786] env[62952]: _type = "Task" [ 854.635786] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.646299] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367102, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.700408] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Releasing lock "refresh_cache-6f085d0c-c440-4bba-9bf8-cf2f70742748" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.700670] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 854.700878] env[62952]: DEBUG nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.701119] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.718880] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.752101] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Releasing lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.752677] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.752981] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89ab4c87-af3d-4e38-8b8a-8b7ff182a0b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.768817] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d92e6f-cf05-47a6-b70c-4ef1c09cad1a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.804956] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 4846f1d8-7628-405e-bef0-0d795a973a34 could not be found. [ 854.805800] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.806234] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dadbcaa4-714c-4eda-b6d7-1eab6cce041e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.817302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521b4751-d1d2-4986-80b5-a6404304aadb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.622932] env[62952]: INFO nova.scheduler.client.report [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted allocations for instance 4be2f0a4-5f57-4a80-9450-6d61d344799e [ 855.631018] env[62952]: DEBUG nova.network.neutron [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.639769] env[62952]: DEBUG oslo_vmware.api [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367102, 'name': PowerOnVM_Task, 'duration_secs': 0.431213} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.641339] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.641540] env[62952]: INFO nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Took 6.38 seconds to spawn the instance on the hypervisor. [ 855.641707] env[62952]: DEBUG nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.642463] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071e3ce0-3bec-41e3-9a61-8e5670cb3a42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.649148] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf6bb79-4282-44e6-8584-094d97b53318 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.652037] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4846f1d8-7628-405e-bef0-0d795a973a34 could not be found. [ 855.652240] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.652402] env[62952]: INFO nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Took 0.90 seconds to destroy the instance on the hypervisor. [ 855.652631] env[62952]: DEBUG oslo.service.loopingcall [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.653204] env[62952]: DEBUG nova.compute.manager [-] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.653293] env[62952]: DEBUG nova.network.neutron [-] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 855.657925] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2d3f00-2eca-46e8-b6df-0b58506bb218 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.692309] env[62952]: DEBUG nova.network.neutron [-] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.693982] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a76688-0da2-4918-b1e0-caa166813984 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.702015] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd9f649-0373-4340-a313-bd396fd8b46d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.717963] env[62952]: DEBUG nova.compute.provider_tree [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.728680] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "7621d6a6-27aa-45d3-80fa-65e957519a1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.728971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.763029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "93951b09-9d7c-42da-83de-0db3980d0d5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.763275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.135425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-affccd29-eef2-4353-a5ae-3184ff82f045 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "4be2f0a4-5f57-4a80-9450-6d61d344799e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 168.783s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.135866] env[62952]: INFO nova.compute.manager [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] [instance: 6f085d0c-c440-4bba-9bf8-cf2f70742748] Took 1.43 seconds to deallocate network for instance. [ 856.154914] env[62952]: DEBUG nova.compute.manager [None req-e89e921c-1717-45d0-b1c9-2857cff4c4a6 tempest-ServerDiagnosticsV248Test-1319815890 tempest-ServerDiagnosticsV248Test-1319815890-project-admin] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.156466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77c75f0-51de-472c-9cce-e3f582d7a617 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.164462] env[62952]: INFO nova.compute.manager [None req-e89e921c-1717-45d0-b1c9-2857cff4c4a6 tempest-ServerDiagnosticsV248Test-1319815890 tempest-ServerDiagnosticsV248Test-1319815890-project-admin] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Retrieving diagnostics [ 856.165264] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692f0500-bef3-4e18-87d6-c516c8ee870a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.173740] env[62952]: INFO nova.compute.manager [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Took 28.85 seconds to build instance. [ 856.199284] env[62952]: DEBUG nova.network.neutron [-] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.202136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-79a48fcc-1aa6-49bb-ba1a-917f01d2345c tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.743s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.220757] env[62952]: DEBUG nova.scheduler.client.report [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.644949] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.704909] env[62952]: INFO nova.compute.manager [-] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Took 1.05 seconds to deallocate network for instance. [ 856.706640] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.725473] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.231s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.725994] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.728840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.825s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.730260] env[62952]: INFO nova.compute.claims [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.036961] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.037203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.165781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.172247] env[62952]: INFO nova.scheduler.client.report [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Deleted allocations for instance 6f085d0c-c440-4bba-9bf8-cf2f70742748 [ 857.229241] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.233855] env[62952]: DEBUG nova.compute.utils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.237531] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.237531] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.262140] env[62952]: INFO nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Took 0.56 seconds to detach 1 volumes for instance. [ 857.264326] env[62952]: DEBUG nova.compute.claims [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 857.264593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.287464] env[62952]: DEBUG nova.policy [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f24a30e1160a43d5be6cde39b8f79edd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '094ca56818604bee9c1f75d1493ab3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.517054] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Successfully created port: 33051a20-0ad0-490d-9c83-c2decd660298 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.683865] env[62952]: DEBUG oslo_concurrency.lockutils [None req-38f8d46c-d83f-4cb0-8f62-cdbdf4a80cf1 tempest-AttachInterfacesUnderV243Test-2115078195 tempest-AttachInterfacesUnderV243Test-2115078195-project-member] Lock "6f085d0c-c440-4bba-9bf8-cf2f70742748" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.426s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.742995] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.121426] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a18230-b434-4e5c-bcbd-95d7e09e93fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.128978] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b824c02-66fc-4bc7-b162-bddd21de6754 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.165393] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa81d37-b23f-48d3-b41c-1746fe876d9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.171392] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836f652b-20de-40e9-ab44-bed35b099cbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.185024] env[62952]: DEBUG nova.compute.provider_tree [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.190178] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.343925] env[62952]: DEBUG nova.compute.manager [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Received event network-changed-33051a20-0ad0-490d-9c83-c2decd660298 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.344184] env[62952]: DEBUG nova.compute.manager [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Refreshing instance network info cache due to event network-changed-33051a20-0ad0-490d-9c83-c2decd660298. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.344628] env[62952]: DEBUG oslo_concurrency.lockutils [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] Acquiring lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.344835] env[62952]: DEBUG oslo_concurrency.lockutils [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] Acquired lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.345055] env[62952]: DEBUG nova.network.neutron [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Refreshing network info cache for port 33051a20-0ad0-490d-9c83-c2decd660298 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.493102] env[62952]: ERROR nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 858.493102] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 858.493102] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 858.493102] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 858.493102] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 858.493102] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 858.493102] env[62952]: ERROR nova.compute.manager raise self.value [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 858.493102] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 858.493102] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 858.493102] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 858.493534] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 858.493534] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 858.493534] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 858.493534] env[62952]: ERROR nova.compute.manager [ 858.493534] env[62952]: Traceback (most recent call last): [ 858.493534] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 858.493534] env[62952]: listener.cb(fileno) [ 858.493534] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 858.493534] env[62952]: result = function(*args, **kwargs) [ 858.493534] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 858.493534] env[62952]: return func(*args, **kwargs) [ 858.493534] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 858.493534] env[62952]: raise e [ 858.493534] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 858.493534] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 858.493534] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 858.493534] env[62952]: created_port_ids = self._update_ports_for_instance( [ 858.493534] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 858.493534] env[62952]: with excutils.save_and_reraise_exception(): [ 858.493534] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 858.493534] env[62952]: self.force_reraise() [ 858.493534] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 858.493534] env[62952]: raise self.value [ 858.493534] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 858.493534] env[62952]: updated_port = self._update_port( [ 858.493534] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 858.493534] env[62952]: _ensure_no_port_binding_failure(port) [ 858.493534] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 858.493534] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 858.494258] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 858.494258] env[62952]: Removing descriptor: 15 [ 858.695515] env[62952]: DEBUG nova.scheduler.client.report [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.720597] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.759961] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.784039] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.784308] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.784469] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.784649] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.784793] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.784937] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.785160] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.785335] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.785503] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.785664] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.785834] env[62952]: DEBUG nova.virt.hardware [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.787046] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade63325-b11c-4c59-b181-9f36be867c04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.795434] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d3f0ae-db1d-4937-962d-c4fc9f0c53c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.810031] env[62952]: ERROR nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Traceback (most recent call last): [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] yield resources [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self.driver.spawn(context, instance, image_meta, [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self._vmops.spawn(context, instance, image_meta, injected_files, [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] vm_ref = self.build_virtual_machine(instance, [ 858.810031] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] vif_infos = vmwarevif.get_vif_info(self._session, [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] for vif in network_info: [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] return self._sync_wrapper(fn, *args, **kwargs) [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self.wait() [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self[:] = self._gt.wait() [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] return self._exit_event.wait() [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 858.810348] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] current.throw(*self._exc) [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] result = function(*args, **kwargs) [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] return func(*args, **kwargs) [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] raise e [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] nwinfo = self.network_api.allocate_for_instance( [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] created_port_ids = self._update_ports_for_instance( [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] with excutils.save_and_reraise_exception(): [ 858.810631] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self.force_reraise() [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] raise self.value [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] updated_port = self._update_port( [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] _ensure_no_port_binding_failure(port) [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] raise exception.PortBindingFailed(port_id=port['id']) [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 858.810988] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] [ 858.810988] env[62952]: INFO nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Terminating instance [ 858.812279] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.875656] env[62952]: DEBUG nova.network.neutron [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.957789] env[62952]: DEBUG nova.network.neutron [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.204131] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.205178] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.210589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.412s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.213225] env[62952]: INFO nova.compute.claims [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.460897] env[62952]: DEBUG oslo_concurrency.lockutils [req-24583466-5c5b-495c-b28a-897a0c801118 req-554b6d54-cd5f-42e8-8a62-0b83da82d612 service nova] Releasing lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.461335] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.461533] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.720774] env[62952]: DEBUG nova.compute.utils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.725387] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.725387] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 859.804031] env[62952]: DEBUG nova.policy [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a7e75f8e6944c27b4a5982f524f3d32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebf2beba58fc43fd8972f670fc3f9c4b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.133606] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.225819] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.289449] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.375315] env[62952]: DEBUG nova.compute.manager [req-ca66380e-d12a-40a7-800b-e04e6e11606f req-77f4d2a9-f591-4e5f-bc25-c23e399e939d service nova] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Received event network-vif-deleted-33051a20-0ad0-490d-9c83-c2decd660298 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.458935] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Successfully created port: 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.558276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3cf041-3cd0-402c-a87f-b037bd6f10c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.565760] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861d3cb5-33bb-4e7a-80c5-7ebfe9ca523b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.596448] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e06ae9-dd31-41b6-831f-d5b07f28dfdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.604305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e204609a-0c4c-4132-b263-d647a3dd6750 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.619311] env[62952]: DEBUG nova.compute.provider_tree [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.792142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.792688] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.792949] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.793318] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd3684cd-3cab-42d7-bc85-329c1a2e0330 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.808185] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2afabaa-0f3b-4357-8648-810142176078 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.833139] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 16224e22-7091-4381-9f40-3f52f24bb724 could not be found. [ 860.833510] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.833829] env[62952]: INFO nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Took 0.04 seconds to destroy the instance on the hypervisor. [ 860.834201] env[62952]: DEBUG oslo.service.loopingcall [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.834517] env[62952]: DEBUG nova.compute.manager [-] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.834703] env[62952]: DEBUG nova.network.neutron [-] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.853861] env[62952]: DEBUG nova.network.neutron [-] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.122734] env[62952]: DEBUG nova.scheduler.client.report [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.239338] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.267938] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.268203] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.268390] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.268533] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.268678] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.268822] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.269382] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.269616] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.269796] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.269964] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.270798] env[62952]: DEBUG nova.virt.hardware [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.271321] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe14c566-ce69-45bb-bf38-fec54535c18f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.279118] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9e82a9-7857-44c3-9710-075673310e05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.358228] env[62952]: DEBUG nova.network.neutron [-] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.465048] env[62952]: ERROR nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 861.465048] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 861.465048] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 861.465048] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 861.465048] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.465048] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.465048] env[62952]: ERROR nova.compute.manager raise self.value [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 861.465048] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 861.465048] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 861.465048] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 861.465508] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 861.465508] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 861.465508] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 861.465508] env[62952]: ERROR nova.compute.manager [ 861.465508] env[62952]: Traceback (most recent call last): [ 861.465508] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 861.465508] env[62952]: listener.cb(fileno) [ 861.465508] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 861.465508] env[62952]: result = function(*args, **kwargs) [ 861.465508] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 861.465508] env[62952]: return func(*args, **kwargs) [ 861.465508] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 861.465508] env[62952]: raise e [ 861.465508] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 861.465508] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 861.465508] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 861.465508] env[62952]: created_port_ids = self._update_ports_for_instance( [ 861.465508] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 861.465508] env[62952]: with excutils.save_and_reraise_exception(): [ 861.465508] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.465508] env[62952]: self.force_reraise() [ 861.465508] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.465508] env[62952]: raise self.value [ 861.465508] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 861.465508] env[62952]: updated_port = self._update_port( [ 861.465508] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 861.465508] env[62952]: _ensure_no_port_binding_failure(port) [ 861.465508] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 861.465508] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 861.466277] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 861.466277] env[62952]: Removing descriptor: 15 [ 861.466277] env[62952]: ERROR nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Traceback (most recent call last): [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] yield resources [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self.driver.spawn(context, instance, image_meta, [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 861.466277] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] vm_ref = self.build_virtual_machine(instance, [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] vif_infos = vmwarevif.get_vif_info(self._session, [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] for vif in network_info: [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return self._sync_wrapper(fn, *args, **kwargs) [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self.wait() [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self[:] = self._gt.wait() [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return self._exit_event.wait() [ 861.466589] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] result = hub.switch() [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return self.greenlet.switch() [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] result = function(*args, **kwargs) [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return func(*args, **kwargs) [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] raise e [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] nwinfo = self.network_api.allocate_for_instance( [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 861.466975] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] created_port_ids = self._update_ports_for_instance( [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] with excutils.save_and_reraise_exception(): [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self.force_reraise() [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] raise self.value [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] updated_port = self._update_port( [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] _ensure_no_port_binding_failure(port) [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 861.467367] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] raise exception.PortBindingFailed(port_id=port['id']) [ 861.467635] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 861.467635] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] [ 861.467635] env[62952]: INFO nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Terminating instance [ 861.468223] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquiring lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.468383] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquired lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.468551] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.630219] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.630747] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.633292] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.762s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.860603] env[62952]: INFO nova.compute.manager [-] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Took 1.03 seconds to deallocate network for instance. [ 861.863305] env[62952]: DEBUG nova.compute.claims [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 861.863305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.991731] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.083444] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.137314] env[62952]: DEBUG nova.compute.utils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.146647] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.149235] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.193273] env[62952]: DEBUG nova.policy [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fda30da3f0e7444b9a109fd15c1bbb20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7a9b722dee442b1bd6fa656a3689fa0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.427494] env[62952]: DEBUG nova.compute.manager [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Received event network-changed-52e46d51-88f2-4c40-911b-ad3a8f7dc6bb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.427494] env[62952]: DEBUG nova.compute.manager [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Refreshing instance network info cache due to event network-changed-52e46d51-88f2-4c40-911b-ad3a8f7dc6bb. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.427494] env[62952]: DEBUG oslo_concurrency.lockutils [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] Acquiring lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.496951] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Successfully created port: aa8d3150-71a3-4f36-9bec-823d2e32b95c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.589094] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Releasing lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.589094] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.589094] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 862.589094] env[62952]: DEBUG oslo_concurrency.lockutils [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] Acquired lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.589094] env[62952]: DEBUG nova.network.neutron [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Refreshing network info cache for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.589252] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5a745d7-d291-4efa-b672-684eafbb0944 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.605647] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb3dab2-2e53-4c51-8c96-a51bc33cda40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.631060] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4ea4cff-ada0-4355-a514-d8b179616535 could not be found. [ 862.632151] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 862.633019] env[62952]: INFO nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Took 0.05 seconds to destroy the instance on the hypervisor. [ 862.633019] env[62952]: DEBUG oslo.service.loopingcall [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.633019] env[62952]: DEBUG nova.compute.manager [-] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.633200] env[62952]: DEBUG nova.network.neutron [-] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 862.649384] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.658726] env[62952]: DEBUG nova.network.neutron [-] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.682248] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 95c07667-e9d8-4cc3-b135-f80d196d179f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.682772] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ca7fc13d-e9be-481c-ae58-78377627800c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.683142] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 34b6c13c-c95b-4cc2-b6eb-cc9374535aec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.683817] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4846f1d8-7628-405e-bef0-0d795a973a34 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.683817] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 16224e22-7091-4381-9f40-3f52f24bb724 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.683817] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance b4ea4cff-ada0-4355-a514-d8b179616535 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 862.684084] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 39948315-6438-48f1-883b-b376e0786650 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.125348] env[62952]: DEBUG nova.network.neutron [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.164215] env[62952]: DEBUG nova.network.neutron [-] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.191018] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 69c93772-822a-4f5b-b2d1-95e82a46f286 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.239919] env[62952]: DEBUG nova.network.neutron [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.649122] env[62952]: ERROR nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 863.649122] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 863.649122] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 863.649122] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 863.649122] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 863.649122] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 863.649122] env[62952]: ERROR nova.compute.manager raise self.value [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 863.649122] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 863.649122] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 863.649122] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 863.649668] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 863.649668] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 863.649668] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 863.649668] env[62952]: ERROR nova.compute.manager [ 863.651653] env[62952]: Traceback (most recent call last): [ 863.651653] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 863.651653] env[62952]: listener.cb(fileno) [ 863.651653] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 863.651653] env[62952]: result = function(*args, **kwargs) [ 863.651653] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 863.651653] env[62952]: return func(*args, **kwargs) [ 863.651653] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 863.651653] env[62952]: raise e [ 863.651653] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 863.651653] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 863.651653] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 863.651653] env[62952]: created_port_ids = self._update_ports_for_instance( [ 863.651653] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 863.651653] env[62952]: with excutils.save_and_reraise_exception(): [ 863.651653] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 863.651653] env[62952]: self.force_reraise() [ 863.651653] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 863.651653] env[62952]: raise self.value [ 863.651653] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 863.651653] env[62952]: updated_port = self._update_port( [ 863.651653] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 863.651653] env[62952]: _ensure_no_port_binding_failure(port) [ 863.651653] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 863.651653] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 863.651653] env[62952]: nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 863.651653] env[62952]: Removing descriptor: 15 [ 863.658699] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.670135] env[62952]: INFO nova.compute.manager [-] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Took 1.04 seconds to deallocate network for instance. [ 863.674019] env[62952]: DEBUG nova.compute.claims [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 863.674019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.686433] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.686667] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.686823] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.687021] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.687173] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.687341] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.687568] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.687731] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.687899] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.688076] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.688252] env[62952]: DEBUG nova.virt.hardware [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.689137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1298a2a-81a2-432f-857a-10f80156c407 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.692584] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 62d47ebb-8782-4062-a27e-556c758ed475 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.701186] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabe0c08-1725-4f3c-aac4-5e26cc63cd6c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.716611] env[62952]: ERROR nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] Traceback (most recent call last): [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] yield resources [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self.driver.spawn(context, instance, image_meta, [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self._vmops.spawn(context, instance, image_meta, injected_files, [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] vm_ref = self.build_virtual_machine(instance, [ 863.716611] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] vif_infos = vmwarevif.get_vif_info(self._session, [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] for vif in network_info: [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] return self._sync_wrapper(fn, *args, **kwargs) [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self.wait() [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self[:] = self._gt.wait() [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] return self._exit_event.wait() [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 863.717162] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] current.throw(*self._exc) [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] result = function(*args, **kwargs) [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] return func(*args, **kwargs) [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] raise e [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] nwinfo = self.network_api.allocate_for_instance( [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] created_port_ids = self._update_ports_for_instance( [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] with excutils.save_and_reraise_exception(): [ 863.717659] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self.force_reraise() [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] raise self.value [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] updated_port = self._update_port( [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] _ensure_no_port_binding_failure(port) [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] raise exception.PortBindingFailed(port_id=port['id']) [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 863.718161] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] [ 863.718161] env[62952]: INFO nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Terminating instance [ 863.720064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.720064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquired lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.720064] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.743520] env[62952]: DEBUG oslo_concurrency.lockutils [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] Releasing lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.743520] env[62952]: DEBUG nova.compute.manager [req-ccec9d34-5763-4fd7-970a-1e546b64e58a req-99a6e2d2-bc89-4988-80f6-2271965c15a2 service nova] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Received event network-vif-deleted-52e46d51-88f2-4c40-911b-ad3a8f7dc6bb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.197942] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 56aa8186-2bcb-47fc-a832-31654cbf218b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.238091] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.320808] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.451029] env[62952]: DEBUG nova.compute.manager [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] [instance: 39948315-6438-48f1-883b-b376e0786650] Received event network-changed-aa8d3150-71a3-4f36-9bec-823d2e32b95c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.452015] env[62952]: DEBUG nova.compute.manager [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] [instance: 39948315-6438-48f1-883b-b376e0786650] Refreshing instance network info cache due to event network-changed-aa8d3150-71a3-4f36-9bec-823d2e32b95c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.452015] env[62952]: DEBUG oslo_concurrency.lockutils [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] Acquiring lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.701402] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 351b1672-ddb3-4562-9965-d65a7d869e82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.824514] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Releasing lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.825044] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.825287] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.825613] env[62952]: DEBUG oslo_concurrency.lockutils [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] Acquired lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.825820] env[62952]: DEBUG nova.network.neutron [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] [instance: 39948315-6438-48f1-883b-b376e0786650] Refreshing network info cache for port aa8d3150-71a3-4f36-9bec-823d2e32b95c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.826917] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a05faa86-0e2e-4e98-b71d-107b79d80dc4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.835455] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2c3723-5503-4ad2-9496-ca70a6dc059d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.857853] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39948315-6438-48f1-883b-b376e0786650 could not be found. [ 864.858072] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.858252] env[62952]: INFO nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Took 0.03 seconds to destroy the instance on the hypervisor. [ 864.858495] env[62952]: DEBUG oslo.service.loopingcall [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.858716] env[62952]: DEBUG nova.compute.manager [-] [instance: 39948315-6438-48f1-883b-b376e0786650] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.858806] env[62952]: DEBUG nova.network.neutron [-] [instance: 39948315-6438-48f1-883b-b376e0786650] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.873722] env[62952]: DEBUG nova.network.neutron [-] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.988107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.988341] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.205048] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance d2faf1e0-cfd1-4d87-ba77-0af92dc16643 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 865.346026] env[62952]: DEBUG nova.network.neutron [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.376394] env[62952]: DEBUG nova.network.neutron [-] [instance: 39948315-6438-48f1-883b-b376e0786650] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.412217] env[62952]: DEBUG nova.network.neutron [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] [instance: 39948315-6438-48f1-883b-b376e0786650] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.709986] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a63ec2ac-4484-4360-962b-105a5338c1b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 865.879013] env[62952]: INFO nova.compute.manager [-] [instance: 39948315-6438-48f1-883b-b376e0786650] Took 1.02 seconds to deallocate network for instance. [ 865.881416] env[62952]: DEBUG nova.compute.claims [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 865.881588] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.914267] env[62952]: DEBUG oslo_concurrency.lockutils [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] Releasing lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.914500] env[62952]: DEBUG nova.compute.manager [req-847e66e0-6d27-4649-bc8e-6a5ede82644b req-11185d6f-2b6f-4369-8e4d-3e51d0d8d285 service nova] [instance: 39948315-6438-48f1-883b-b376e0786650] Received event network-vif-deleted-aa8d3150-71a3-4f36-9bec-823d2e32b95c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.214981] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 802fd770-38aa-4417-b25e-fd83510bceef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 866.697407] env[62952]: DEBUG nova.compute.manager [None req-65a6d154-df4e-4e56-bc97-f512ab3316cf tempest-ServerDiagnosticsV248Test-1319815890 tempest-ServerDiagnosticsV248Test-1319815890-project-admin] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.698740] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac4a82a-03e7-4062-a120-402ae8781c0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.705965] env[62952]: INFO nova.compute.manager [None req-65a6d154-df4e-4e56-bc97-f512ab3316cf tempest-ServerDiagnosticsV248Test-1319815890 tempest-ServerDiagnosticsV248Test-1319815890-project-admin] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Retrieving diagnostics [ 866.706664] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fedb15-c85f-44a2-a65c-41a40481a570 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.733921] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 2f475040-28df-47e4-bfc4-4c843a57885f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.237841] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.741167] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 5914a35a-0934-4f8f-81e8-d91bc690a9cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.762877] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.763159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.763367] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.763549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.763717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.766550] env[62952]: INFO nova.compute.manager [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Terminating instance [ 867.770615] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "refresh_cache-34b6c13c-c95b-4cc2-b6eb-cc9374535aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.770837] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquired lock "refresh_cache-34b6c13c-c95b-4cc2-b6eb-cc9374535aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.771031] env[62952]: DEBUG nova.network.neutron [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.244776] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 34f5995f-692c-4f0d-8b15-0d388df7e34e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 868.288646] env[62952]: DEBUG nova.network.neutron [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.352072] env[62952]: DEBUG nova.network.neutron [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.748058] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4983f354-77dc-4e84-969b-4c74441fa568 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 868.855214] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Releasing lock "refresh_cache-34b6c13c-c95b-4cc2-b6eb-cc9374535aec" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.855659] env[62952]: DEBUG nova.compute.manager [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.855854] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.856748] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eabe19-dfe2-4ee9-9c24-c1bb79c3ee08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.864584] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.864829] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45488528-66f0-41e7-bb46-f860bdbfc78e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.871348] env[62952]: DEBUG oslo_vmware.api [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 868.871348] env[62952]: value = "task-1367103" [ 868.871348] env[62952]: _type = "Task" [ 868.871348] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.879732] env[62952]: DEBUG oslo_vmware.api [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.253022] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 7621d6a6-27aa-45d3-80fa-65e957519a1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 869.381362] env[62952]: DEBUG oslo_vmware.api [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367103, 'name': PowerOffVM_Task, 'duration_secs': 0.116118} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.381626] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.381792] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.382043] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c516e0c-e5c5-4918-94bc-ee4e7a8fcf91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.407957] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.408199] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.408376] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Deleting the datastore file [datastore2] 34b6c13c-c95b-4cc2-b6eb-cc9374535aec {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.408636] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-698e2375-81e3-40c6-9aee-209c7f5f86df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.416040] env[62952]: DEBUG oslo_vmware.api [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for the task: (returnval){ [ 869.416040] env[62952]: value = "task-1367105" [ 869.416040] env[62952]: _type = "Task" [ 869.416040] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.423919] env[62952]: DEBUG oslo_vmware.api [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367105, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.755875] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 93951b09-9d7c-42da-83de-0db3980d0d5d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 869.926558] env[62952]: DEBUG oslo_vmware.api [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Task: {'id': task-1367105, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10012} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.926558] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.926816] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.926816] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.927047] env[62952]: INFO nova.compute.manager [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Took 1.07 seconds to destroy the instance on the hypervisor. [ 869.927294] env[62952]: DEBUG oslo.service.loopingcall [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.927536] env[62952]: DEBUG nova.compute.manager [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.927679] env[62952]: DEBUG nova.network.neutron [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 869.942739] env[62952]: DEBUG nova.network.neutron [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.258856] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 94fd5287-9a8e-45e4-99e8-0bcc861f889c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 870.259185] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 870.259629] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 870.445688] env[62952]: DEBUG nova.network.neutron [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.517948] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f1c997-e379-4903-97c8-c7fa8783c530 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.525522] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54e3083-8733-4cf6-89c0-4640698b27ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.555357] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb4ce8e-5297-464c-bb46-dbb7b86e1773 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.562213] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38b5a11-3d6d-40f6-9909-f49ef29ed967 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.575210] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.949729] env[62952]: INFO nova.compute.manager [-] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Took 1.02 seconds to deallocate network for instance. [ 871.078741] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.458178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.584249] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 871.584502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.951s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.584775] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.810s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.587609] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 871.587757] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Cleaning up deleted instances {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 872.093572] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] There are 2 instances to clean {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 872.093837] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9567865d-41ae-4379-8313-e87bfc92fba8] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 872.328705] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2ba0d2-df0e-41ef-acc8-07f0a37a4ff4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.336463] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e382f227-24ad-44b5-8d0c-0d523e068a4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.366273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741619f7-8c49-44a3-9572-5bc4cd5977ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.373278] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9300e3-350d-4871-8083-fa798d2d48f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.385830] env[62952]: DEBUG nova.compute.provider_tree [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.597572] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 418cd947-00b8-485f-b9fc-2a675c58998d] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 872.891173] env[62952]: DEBUG nova.scheduler.client.report [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.101671] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.101850] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Cleaning up deleted instances with incomplete migration {{(pid=62952) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 873.394810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.809s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.395177] env[62952]: ERROR nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Traceback (most recent call last): [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self.driver.spawn(context, instance, image_meta, [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] vm_ref = self.build_virtual_machine(instance, [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] vif_infos = vmwarevif.get_vif_info(self._session, [ 873.395177] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] for vif in network_info: [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return self._sync_wrapper(fn, *args, **kwargs) [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self.wait() [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self[:] = self._gt.wait() [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return self._exit_event.wait() [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] result = hub.switch() [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 873.395504] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return self.greenlet.switch() [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] result = function(*args, **kwargs) [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] return func(*args, **kwargs) [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] raise e [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] nwinfo = self.network_api.allocate_for_instance( [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] created_port_ids = self._update_ports_for_instance( [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] with excutils.save_and_reraise_exception(): [ 873.395838] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] self.force_reraise() [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] raise self.value [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] updated_port = self._update_port( [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] _ensure_no_port_binding_failure(port) [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] raise exception.PortBindingFailed(port_id=port['id']) [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] nova.exception.PortBindingFailed: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. [ 873.396182] env[62952]: ERROR nova.compute.manager [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] [ 873.396462] env[62952]: DEBUG nova.compute.utils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 873.398109] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Build of instance 95c07667-e9d8-4cc3-b135-f80d196d179f was re-scheduled: Binding failed for port af77a34b-4d47-42de-97f3-b3afc5627973, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 873.398568] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 873.398848] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquiring lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.399057] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Acquired lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.399267] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.400280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.215s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.604502] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.923129] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.017910] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.163433] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39df584-c1a3-41db-9573-3ca63cfd0aa1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.170321] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5ea553-d7bb-494d-9a3f-55ec5e0f5fef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.199599] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e793cea7-6564-43b4-95d5-49b1b116830e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.206378] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255da025-a4be-4f6f-a57c-d8a21feae552 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.218929] env[62952]: DEBUG nova.compute.provider_tree [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.520114] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Releasing lock "refresh_cache-95c07667-e9d8-4cc3-b135-f80d196d179f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.520385] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 874.520574] env[62952]: DEBUG nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.520742] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.535461] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.722608] env[62952]: DEBUG nova.scheduler.client.report [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.038547] env[62952]: DEBUG nova.network.neutron [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.227875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.827s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.228562] env[62952]: ERROR nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Traceback (most recent call last): [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self.driver.spawn(context, instance, image_meta, [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] vm_ref = self.build_virtual_machine(instance, [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] vif_infos = vmwarevif.get_vif_info(self._session, [ 875.228562] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] for vif in network_info: [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return self._sync_wrapper(fn, *args, **kwargs) [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self.wait() [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self[:] = self._gt.wait() [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return self._exit_event.wait() [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] result = hub.switch() [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 875.228896] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return self.greenlet.switch() [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] result = function(*args, **kwargs) [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] return func(*args, **kwargs) [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] raise e [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] nwinfo = self.network_api.allocate_for_instance( [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] created_port_ids = self._update_ports_for_instance( [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] with excutils.save_and_reraise_exception(): [ 875.229199] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] self.force_reraise() [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] raise self.value [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] updated_port = self._update_port( [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] _ensure_no_port_binding_failure(port) [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] raise exception.PortBindingFailed(port_id=port['id']) [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] nova.exception.PortBindingFailed: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. [ 875.229533] env[62952]: ERROR nova.compute.manager [instance: ca7fc13d-e9be-481c-ae58-78377627800c] [ 875.229834] env[62952]: DEBUG nova.compute.utils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 875.230551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.540s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.232030] env[62952]: INFO nova.compute.claims [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.235237] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Build of instance ca7fc13d-e9be-481c-ae58-78377627800c was re-scheduled: Binding failed for port d90be0de-44f3-429b-831a-f29cb3a5182a, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 875.235674] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 875.235897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquiring lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.236055] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Acquired lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.236244] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.541425] env[62952]: INFO nova.compute.manager [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] [instance: 95c07667-e9d8-4cc3-b135-f80d196d179f] Took 1.02 seconds to deallocate network for instance. [ 875.755541] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.808599] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.311421] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Releasing lock "refresh_cache-ca7fc13d-e9be-481c-ae58-78377627800c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.312195] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 876.312195] env[62952]: DEBUG nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.312195] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 876.326320] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.484392] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474d4396-4b66-4d0c-953f-0fbc8c352c34 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.491792] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f8690c-1b56-447d-8ff4-4992df160b35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.519894] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf20cae-9932-47b6-b5b8-31295add1ea3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.526448] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f781576-b016-45e2-bbde-2449bd0873ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.538789] env[62952]: DEBUG nova.compute.provider_tree [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.566960] env[62952]: INFO nova.scheduler.client.report [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Deleted allocations for instance 95c07667-e9d8-4cc3-b135-f80d196d179f [ 876.828840] env[62952]: DEBUG nova.network.neutron [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.042087] env[62952]: DEBUG nova.scheduler.client.report [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.077233] env[62952]: DEBUG oslo_concurrency.lockutils [None req-712b39f9-2797-46fa-8e8c-140fb2669353 tempest-VolumesAdminNegativeTest-534090508 tempest-VolumesAdminNegativeTest-534090508-project-member] Lock "95c07667-e9d8-4cc3-b135-f80d196d179f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 174.534s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.331770] env[62952]: INFO nova.compute.manager [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] [instance: ca7fc13d-e9be-481c-ae58-78377627800c] Took 1.02 seconds to deallocate network for instance. [ 877.546754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.547305] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.549850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.384s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.551346] env[62952]: INFO nova.compute.claims [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.579530] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 878.058467] env[62952]: DEBUG nova.compute.utils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.063539] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 878.063725] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 878.104374] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.143803] env[62952]: DEBUG nova.policy [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '051f2e5a76a941f38567bd7fead72aa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '069d0ad1561543668252ccb60e2c33c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.364547] env[62952]: INFO nova.scheduler.client.report [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Deleted allocations for instance ca7fc13d-e9be-481c-ae58-78377627800c [ 878.564997] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.582648] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Successfully created port: 3951bf5b-4bf7-4759-acac-0c83678ccf69 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.865680] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77470b32-4394-42da-8049-b0b6cfece261 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.874064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7190c579-7a97-4b2f-938b-9d9750e8a2cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.878567] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8771530a-417d-4d7b-aa40-87c2fead12f6 tempest-TenantUsagesTestJSON-1176666093 tempest-TenantUsagesTestJSON-1176666093-project-member] Lock "ca7fc13d-e9be-481c-ae58-78377627800c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.034s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.915029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532c9682-c04b-4593-8449-d3bb4816aeb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.923176] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a8ca8e-feae-4be4-8cd8-114da55eefd1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.938354] env[62952]: DEBUG nova.compute.provider_tree [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.382058] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.441690] env[62952]: DEBUG nova.scheduler.client.report [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.577183] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.607885] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.607885] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.607885] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.608057] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.608057] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.608928] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.608928] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.608928] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.609358] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.609668] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.609925] env[62952]: DEBUG nova.virt.hardware [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.611231] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecf76cd-cc38-4b70-856d-9656d5d57fd8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.621235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286b91bf-26e3-445d-b295-ea9b1eebb273 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.631637] env[62952]: DEBUG nova.compute.manager [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Received event network-changed-3951bf5b-4bf7-4759-acac-0c83678ccf69 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.631637] env[62952]: DEBUG nova.compute.manager [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Refreshing instance network info cache due to event network-changed-3951bf5b-4bf7-4759-acac-0c83678ccf69. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.632768] env[62952]: DEBUG oslo_concurrency.lockutils [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] Acquiring lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.632768] env[62952]: DEBUG oslo_concurrency.lockutils [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] Acquired lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.632768] env[62952]: DEBUG nova.network.neutron [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Refreshing network info cache for port 3951bf5b-4bf7-4759-acac-0c83678ccf69 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.741375] env[62952]: ERROR nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 879.741375] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 879.741375] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 879.741375] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 879.741375] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 879.741375] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 879.741375] env[62952]: ERROR nova.compute.manager raise self.value [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 879.741375] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 879.741375] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 879.741375] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 879.741884] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 879.741884] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 879.741884] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 879.741884] env[62952]: ERROR nova.compute.manager [ 879.743870] env[62952]: Traceback (most recent call last): [ 879.743912] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 879.743912] env[62952]: listener.cb(fileno) [ 879.743912] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 879.743912] env[62952]: result = function(*args, **kwargs) [ 879.743912] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 879.743912] env[62952]: return func(*args, **kwargs) [ 879.743912] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 879.743912] env[62952]: raise e [ 879.743912] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 879.743912] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 879.744173] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 879.744173] env[62952]: created_port_ids = self._update_ports_for_instance( [ 879.744173] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 879.744173] env[62952]: with excutils.save_and_reraise_exception(): [ 879.744173] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 879.744173] env[62952]: self.force_reraise() [ 879.744173] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 879.744173] env[62952]: raise self.value [ 879.744173] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 879.744173] env[62952]: updated_port = self._update_port( [ 879.744173] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 879.744173] env[62952]: _ensure_no_port_binding_failure(port) [ 879.744173] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 879.744173] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 879.744173] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 879.744173] env[62952]: Removing descriptor: 15 [ 879.744960] env[62952]: ERROR nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Traceback (most recent call last): [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] yield resources [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self.driver.spawn(context, instance, image_meta, [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self._vmops.spawn(context, instance, image_meta, injected_files, [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] vm_ref = self.build_virtual_machine(instance, [ 879.744960] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] vif_infos = vmwarevif.get_vif_info(self._session, [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] for vif in network_info: [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return self._sync_wrapper(fn, *args, **kwargs) [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self.wait() [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self[:] = self._gt.wait() [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return self._exit_event.wait() [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 879.745261] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] result = hub.switch() [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return self.greenlet.switch() [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] result = function(*args, **kwargs) [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return func(*args, **kwargs) [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] raise e [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] nwinfo = self.network_api.allocate_for_instance( [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] created_port_ids = self._update_ports_for_instance( [ 879.745572] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] with excutils.save_and_reraise_exception(): [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self.force_reraise() [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] raise self.value [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] updated_port = self._update_port( [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] _ensure_no_port_binding_failure(port) [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] raise exception.PortBindingFailed(port_id=port['id']) [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 879.745886] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] [ 879.746211] env[62952]: INFO nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Terminating instance [ 879.747663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquiring lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.904042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.950017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.950017] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.958018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.723s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.958018] env[62952]: INFO nova.compute.claims [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.164666] env[62952]: DEBUG nova.network.neutron [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.285207] env[62952]: DEBUG nova.network.neutron [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.466022] env[62952]: DEBUG nova.compute.utils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.472839] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.473026] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.524235] env[62952]: DEBUG nova.policy [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb541b4abd9426ea96edcd066c69be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d1f373267464f14ae2c833151821973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.791333] env[62952]: DEBUG oslo_concurrency.lockutils [req-7b97df91-90da-4243-91ce-671074affc2d req-7745c018-96a0-4b78-8fc4-bcea8d699e79 service nova] Releasing lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.791333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquired lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.791333] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.815053] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Successfully created port: fe3a90cb-d12d-4ef6-a2fa-c058530e56a3 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.980349] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.310278] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6b3b61-e919-4237-8e28-463447d411e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.318354] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.321662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c457e3b-29a6-4b3d-bc63-a9bbe15982ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.352248] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537a6ba0-8b8e-491a-90a2-c81548b4746f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.359684] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00248ad9-a186-48de-9ff8-790f55658fbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.373396] env[62952]: DEBUG nova.compute.provider_tree [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.481974] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.693334] env[62952]: DEBUG nova.compute.manager [req-688d2a77-4810-423b-94dd-720376756af5 req-aac33653-0818-47f2-a645-ec4c42549f77 service nova] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Received event network-vif-deleted-3951bf5b-4bf7-4759-acac-0c83678ccf69 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.881836] env[62952]: DEBUG nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.984486] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Releasing lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.985052] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.985385] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.985760] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4116a8d-80d7-4ab8-a190-59187c5a73ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.992921] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.998270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160b35bd-48fe-4ea8-8a7a-3941314a4e5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.017083] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.017322] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.017502] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.017684] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.017828] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.017974] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.021639] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.021812] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.022091] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.025741] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.025741] env[62952]: DEBUG nova.virt.hardware [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.025741] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda5d96d-fe47-43b9-bc02-e7bbd6b2cc05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.030316] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69c93772-822a-4f5b-b2d1-95e82a46f286 could not be found. [ 882.030581] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.031055] env[62952]: INFO nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Took 0.05 seconds to destroy the instance on the hypervisor. [ 882.031139] env[62952]: DEBUG oslo.service.loopingcall [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.031926] env[62952]: DEBUG nova.compute.manager [-] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.032089] env[62952]: DEBUG nova.network.neutron [-] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.038496] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7033b310-0469-42c2-aa2a-ad1a416d2bb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.044605] env[62952]: ERROR nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 882.044605] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 882.044605] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 882.044605] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 882.044605] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 882.044605] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 882.044605] env[62952]: ERROR nova.compute.manager raise self.value [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 882.044605] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 882.044605] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 882.044605] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 882.045100] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 882.045100] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 882.045100] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 882.045100] env[62952]: ERROR nova.compute.manager [ 882.045100] env[62952]: Traceback (most recent call last): [ 882.045100] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 882.045100] env[62952]: listener.cb(fileno) [ 882.045100] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 882.045100] env[62952]: result = function(*args, **kwargs) [ 882.045100] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 882.045100] env[62952]: return func(*args, **kwargs) [ 882.045100] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 882.045100] env[62952]: raise e [ 882.045100] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 882.045100] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 882.045100] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 882.045100] env[62952]: created_port_ids = self._update_ports_for_instance( [ 882.045100] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 882.045100] env[62952]: with excutils.save_and_reraise_exception(): [ 882.045100] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 882.045100] env[62952]: self.force_reraise() [ 882.045100] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 882.045100] env[62952]: raise self.value [ 882.045100] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 882.045100] env[62952]: updated_port = self._update_port( [ 882.045100] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 882.045100] env[62952]: _ensure_no_port_binding_failure(port) [ 882.045100] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 882.045100] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 882.045799] env[62952]: nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 882.045799] env[62952]: Removing descriptor: 15 [ 882.056100] env[62952]: ERROR nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Traceback (most recent call last): [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] yield resources [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self.driver.spawn(context, instance, image_meta, [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self._vmops.spawn(context, instance, image_meta, injected_files, [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] vm_ref = self.build_virtual_machine(instance, [ 882.056100] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] vif_infos = vmwarevif.get_vif_info(self._session, [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] for vif in network_info: [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] return self._sync_wrapper(fn, *args, **kwargs) [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self.wait() [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self[:] = self._gt.wait() [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] return self._exit_event.wait() [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 882.056436] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] current.throw(*self._exc) [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] result = function(*args, **kwargs) [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] return func(*args, **kwargs) [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] raise e [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] nwinfo = self.network_api.allocate_for_instance( [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] created_port_ids = self._update_ports_for_instance( [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] with excutils.save_and_reraise_exception(): [ 882.056859] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self.force_reraise() [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] raise self.value [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] updated_port = self._update_port( [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] _ensure_no_port_binding_failure(port) [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] raise exception.PortBindingFailed(port_id=port['id']) [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 882.057208] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] [ 882.057208] env[62952]: INFO nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Terminating instance [ 882.059029] env[62952]: DEBUG nova.network.neutron [-] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.060353] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.060572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.060757] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.387509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.388063] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.392231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.128s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.562672] env[62952]: DEBUG nova.network.neutron [-] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.584283] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.717028] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.902158] env[62952]: DEBUG nova.compute.utils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.905943] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.907361] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.000867] env[62952]: DEBUG nova.policy [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68efb83d9ad14321a8bc7a37b48dd2b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba4ad19b3f434439bda6909ec0836a09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.072422] env[62952]: INFO nova.compute.manager [-] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Took 1.04 seconds to deallocate network for instance. [ 883.081794] env[62952]: DEBUG nova.compute.claims [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 883.082049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.219786] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.220267] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.220497] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.221274] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f4678ae-d05d-4492-9b38-80d951e874e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.231604] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414c1329-668c-42b8-bd93-0b2722c0886e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.261290] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 62d47ebb-8782-4062-a27e-556c758ed475 could not be found. [ 883.261604] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.261969] env[62952]: INFO nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Took 0.04 seconds to destroy the instance on the hypervisor. [ 883.262196] env[62952]: DEBUG oslo.service.loopingcall [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.267252] env[62952]: DEBUG nova.compute.manager [-] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.267252] env[62952]: DEBUG nova.network.neutron [-] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 883.299084] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f071d0aa-bd6f-4987-a78a-5408fc0b59fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.307086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75950577-829b-4728-bbec-3353423d7a61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.341374] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf561222-e3a3-468e-8508-11ff4b439075 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.348718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4349f8-41f8-45ca-a937-4c5d91a416ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.363168] env[62952]: DEBUG nova.compute.provider_tree [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.406412] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.458373] env[62952]: DEBUG nova.network.neutron [-] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 883.705805] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Successfully created port: 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.758405] env[62952]: DEBUG nova.compute.manager [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Received event network-changed-fe3a90cb-d12d-4ef6-a2fa-c058530e56a3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.758405] env[62952]: DEBUG nova.compute.manager [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Refreshing instance network info cache due to event network-changed-fe3a90cb-d12d-4ef6-a2fa-c058530e56a3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.758405] env[62952]: DEBUG oslo_concurrency.lockutils [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] Acquiring lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.758405] env[62952]: DEBUG oslo_concurrency.lockutils [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] Acquired lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.758405] env[62952]: DEBUG nova.network.neutron [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Refreshing network info cache for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.869224] env[62952]: DEBUG nova.scheduler.client.report [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.961479] env[62952]: DEBUG nova.network.neutron [-] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.280060] env[62952]: DEBUG nova.network.neutron [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.380708] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.380928] env[62952]: ERROR nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Traceback (most recent call last): [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self.driver.spawn(context, instance, image_meta, [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self._vmops.spawn(context, instance, image_meta, injected_files, [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] vm_ref = self.build_virtual_machine(instance, [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] vif_infos = vmwarevif.get_vif_info(self._session, [ 884.380928] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] for vif in network_info: [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] return self._sync_wrapper(fn, *args, **kwargs) [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self.wait() [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self[:] = self._gt.wait() [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] return self._exit_event.wait() [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] current.throw(*self._exc) [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 884.381522] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] result = function(*args, **kwargs) [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] return func(*args, **kwargs) [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] raise e [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] nwinfo = self.network_api.allocate_for_instance( [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] created_port_ids = self._update_ports_for_instance( [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] with excutils.save_and_reraise_exception(): [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] self.force_reraise() [ 884.381835] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] raise self.value [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] updated_port = self._update_port( [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] _ensure_no_port_binding_failure(port) [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] raise exception.PortBindingFailed(port_id=port['id']) [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] nova.exception.PortBindingFailed: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. [ 884.382134] env[62952]: ERROR nova.compute.manager [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] [ 884.382134] env[62952]: DEBUG nova.compute.utils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 884.383452] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Build of instance 4846f1d8-7628-405e-bef0-0d795a973a34 was re-scheduled: Binding failed for port a70da931-6354-49a6-aa2c-7d085de6ae67, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 884.384265] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 884.384265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquiring lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.384826] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Acquired lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.385162] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.388068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.666s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.389065] env[62952]: INFO nova.compute.claims [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.420033] env[62952]: DEBUG nova.network.neutron [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.421725] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.455956] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.456251] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.456536] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.456575] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.456719] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.456866] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.457089] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.457284] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.457425] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.457585] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.457756] env[62952]: DEBUG nova.virt.hardware [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.458643] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21726f1b-4788-4ed0-bd9c-e88fa0c392f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.466962] env[62952]: INFO nova.compute.manager [-] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Took 1.20 seconds to deallocate network for instance. [ 884.469704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ece2d51-3dda-42b2-965b-ba1a712a34ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.475769] env[62952]: DEBUG nova.compute.claims [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 884.475969] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.930201] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.930201] env[62952]: DEBUG oslo_concurrency.lockutils [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] Releasing lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.930201] env[62952]: DEBUG nova.compute.manager [req-7facb71c-2593-45e8-8c18-0d2311106ba9 req-564d83b4-1c4b-4a48-b59b-80acf534cf5d service nova] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Received event network-vif-deleted-fe3a90cb-d12d-4ef6-a2fa-c058530e56a3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.117925] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.408617] env[62952]: ERROR nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 885.408617] env[62952]: ERROR nova.compute.manager Traceback (most recent call last): [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 885.408617] env[62952]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 885.408617] env[62952]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 885.408617] env[62952]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 885.408617] env[62952]: ERROR nova.compute.manager self.force_reraise() [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 885.408617] env[62952]: ERROR nova.compute.manager raise self.value [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 885.408617] env[62952]: ERROR nova.compute.manager updated_port = self._update_port( [ 885.408617] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 885.408617] env[62952]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 885.409262] env[62952]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 885.409262] env[62952]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 885.409262] env[62952]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 885.409262] env[62952]: ERROR nova.compute.manager [ 885.409262] env[62952]: Traceback (most recent call last): [ 885.409262] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 885.409262] env[62952]: listener.cb(fileno) [ 885.409262] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 885.409262] env[62952]: result = function(*args, **kwargs) [ 885.409262] env[62952]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 885.409262] env[62952]: return func(*args, **kwargs) [ 885.409262] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 885.409262] env[62952]: raise e [ 885.409262] env[62952]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 885.409262] env[62952]: nwinfo = self.network_api.allocate_for_instance( [ 885.409262] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 885.409262] env[62952]: created_port_ids = self._update_ports_for_instance( [ 885.409262] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 885.409262] env[62952]: with excutils.save_and_reraise_exception(): [ 885.409262] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 885.409262] env[62952]: self.force_reraise() [ 885.409262] env[62952]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 885.409262] env[62952]: raise self.value [ 885.409262] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 885.409262] env[62952]: updated_port = self._update_port( [ 885.409262] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 885.409262] env[62952]: _ensure_no_port_binding_failure(port) [ 885.409262] env[62952]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 885.409262] env[62952]: raise exception.PortBindingFailed(port_id=port['id']) [ 885.410035] env[62952]: nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 885.410035] env[62952]: Removing descriptor: 19 [ 885.410035] env[62952]: ERROR nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Traceback (most recent call last): [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] yield resources [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self.driver.spawn(context, instance, image_meta, [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 885.410035] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] vm_ref = self.build_virtual_machine(instance, [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] vif_infos = vmwarevif.get_vif_info(self._session, [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] for vif in network_info: [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return self._sync_wrapper(fn, *args, **kwargs) [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self.wait() [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self[:] = self._gt.wait() [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return self._exit_event.wait() [ 885.410402] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] result = hub.switch() [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return self.greenlet.switch() [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] result = function(*args, **kwargs) [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return func(*args, **kwargs) [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] raise e [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] nwinfo = self.network_api.allocate_for_instance( [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 885.410928] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] created_port_ids = self._update_ports_for_instance( [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] with excutils.save_and_reraise_exception(): [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self.force_reraise() [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] raise self.value [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] updated_port = self._update_port( [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] _ensure_no_port_binding_failure(port) [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 885.411346] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] raise exception.PortBindingFailed(port_id=port['id']) [ 885.411661] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 885.411661] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] [ 885.411661] env[62952]: INFO nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Terminating instance [ 885.415333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquiring lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.415333] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquired lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.415333] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.622152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Releasing lock "refresh_cache-4846f1d8-7628-405e-bef0-0d795a973a34" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.622152] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 885.622339] env[62952]: DEBUG nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.622501] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.641771] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.723146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec0dda2-2427-4479-84d0-e5cdf977d66f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.732690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48af90dd-c401-495b-8aa0-53c8629d9c5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.773845] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07664e59-d99e-405f-8db5-dd8a002a727a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.781931] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38690db9-7701-4ea7-8cc1-0ac382e968fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.795208] env[62952]: DEBUG nova.compute.provider_tree [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.850433] env[62952]: DEBUG nova.compute.manager [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Received event network-changed-1f6c79c1-d3bd-4eb2-971c-0e6da8192f65 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.850433] env[62952]: DEBUG nova.compute.manager [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Refreshing instance network info cache due to event network-changed-1f6c79c1-d3bd-4eb2-971c-0e6da8192f65. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.850433] env[62952]: DEBUG oslo_concurrency.lockutils [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] Acquiring lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.935999] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.144777] env[62952]: DEBUG nova.network.neutron [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.242602] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.298288] env[62952]: DEBUG nova.scheduler.client.report [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.647682] env[62952]: INFO nova.compute.manager [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] [instance: 4846f1d8-7628-405e-bef0-0d795a973a34] Took 1.02 seconds to deallocate network for instance. [ 886.746659] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Releasing lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.747108] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 886.747464] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.747653] env[62952]: DEBUG oslo_concurrency.lockutils [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] Acquired lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.747836] env[62952]: DEBUG nova.network.neutron [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Refreshing network info cache for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.748883] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4ae143c-e868-406b-b52c-170bc48abeb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.758189] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85611ea-b9af-4b1b-b992-451939ccddd7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.780555] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 56aa8186-2bcb-47fc-a832-31654cbf218b could not be found. [ 886.780555] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.780555] env[62952]: INFO nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 886.780796] env[62952]: DEBUG oslo.service.loopingcall [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.781893] env[62952]: DEBUG nova.compute.manager [-] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.781893] env[62952]: DEBUG nova.network.neutron [-] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.800428] env[62952]: DEBUG nova.network.neutron [-] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.804517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.805014] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.809147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.946s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.271639] env[62952]: DEBUG nova.network.neutron [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.310764] env[62952]: DEBUG nova.network.neutron [-] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.312888] env[62952]: DEBUG nova.compute.utils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.318153] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.318153] env[62952]: DEBUG nova.network.neutron [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 887.393799] env[62952]: DEBUG nova.policy [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c27b33f60824d32862b29f1bd938f55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69a5b81b61794ec0a8d884692ec0ed01', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 887.426368] env[62952]: DEBUG nova.network.neutron [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.620912] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3256bf9a-970d-4535-9ca0-002d81575575 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.627018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df70904c-8573-477c-9c3e-375a068b2472 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.655834] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa42f92b-8d9c-455b-a41a-7dbb03c26e4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.669072] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa3e72c-b106-4163-a30d-691c370e0eb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.684363] env[62952]: DEBUG nova.compute.provider_tree [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.684363] env[62952]: INFO nova.scheduler.client.report [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Deleted allocations for instance 4846f1d8-7628-405e-bef0-0d795a973a34 [ 887.693023] env[62952]: DEBUG nova.network.neutron [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Successfully created port: 06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.814409] env[62952]: INFO nova.compute.manager [-] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Took 1.03 seconds to deallocate network for instance. [ 887.816050] env[62952]: DEBUG nova.compute.claims [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Aborting claim: {{(pid=62952) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 887.816616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.823220] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.933141] env[62952]: DEBUG oslo_concurrency.lockutils [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] Releasing lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.933141] env[62952]: DEBUG nova.compute.manager [req-939557b9-04df-4abd-b33b-a2656bc263f5 req-93bd93f9-47cd-433d-b62f-3e25aece4b5a service nova] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Received event network-vif-deleted-1f6c79c1-d3bd-4eb2-971c-0e6da8192f65 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.188193] env[62952]: DEBUG nova.scheduler.client.report [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.191576] env[62952]: DEBUG oslo_concurrency.lockutils [None req-19d15424-982e-4454-97dc-18c7825d4208 tempest-ServersTestBootFromVolume-52030683 tempest-ServersTestBootFromVolume-52030683-project-member] Lock "4846f1d8-7628-405e-bef0-0d795a973a34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.468s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.694938] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.701446] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.701446] env[62952]: ERROR nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Traceback (most recent call last): [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self.driver.spawn(context, instance, image_meta, [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self._vmops.spawn(context, instance, image_meta, injected_files, [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 888.701446] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] vm_ref = self.build_virtual_machine(instance, [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] vif_infos = vmwarevif.get_vif_info(self._session, [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] for vif in network_info: [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] return self._sync_wrapper(fn, *args, **kwargs) [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self.wait() [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self[:] = self._gt.wait() [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] return self._exit_event.wait() [ 888.701854] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] current.throw(*self._exc) [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] result = function(*args, **kwargs) [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] return func(*args, **kwargs) [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] raise e [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] nwinfo = self.network_api.allocate_for_instance( [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] created_port_ids = self._update_ports_for_instance( [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 888.702227] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] with excutils.save_and_reraise_exception(): [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] self.force_reraise() [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] raise self.value [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] updated_port = self._update_port( [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] _ensure_no_port_binding_failure(port) [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] raise exception.PortBindingFailed(port_id=port['id']) [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] nova.exception.PortBindingFailed: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. [ 888.702606] env[62952]: ERROR nova.compute.manager [instance: 16224e22-7091-4381-9f40-3f52f24bb724] [ 888.702929] env[62952]: DEBUG nova.compute.utils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 888.706338] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Build of instance 16224e22-7091-4381-9f40-3f52f24bb724 was re-scheduled: Binding failed for port 33051a20-0ad0-490d-9c83-c2decd660298, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 888.707883] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 888.707883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.707883] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.707883] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.709334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.036s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.832920] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 888.859368] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.859599] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.859751] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.859980] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.860083] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.860229] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.860430] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.860598] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.860763] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.860945] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.861163] env[62952]: DEBUG nova.virt.hardware [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.862076] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df889409-af29-49bb-b8cb-dfeb4027eb67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.871595] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb847b4-c88f-41a4-aaab-d34fb19dd4fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.236790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.240088] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.472911] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.480813] env[62952]: DEBUG nova.compute.manager [req-ef874528-6b1f-4a08-b2be-809a9889620f req-8bced861-652d-4523-968e-470e458361f0 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Received event network-vif-plugged-06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.480863] env[62952]: DEBUG oslo_concurrency.lockutils [req-ef874528-6b1f-4a08-b2be-809a9889620f req-8bced861-652d-4523-968e-470e458361f0 service nova] Acquiring lock "351b1672-ddb3-4562-9965-d65a7d869e82-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.481392] env[62952]: DEBUG oslo_concurrency.lockutils [req-ef874528-6b1f-4a08-b2be-809a9889620f req-8bced861-652d-4523-968e-470e458361f0 service nova] Lock "351b1672-ddb3-4562-9965-d65a7d869e82-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.481392] env[62952]: DEBUG oslo_concurrency.lockutils [req-ef874528-6b1f-4a08-b2be-809a9889620f req-8bced861-652d-4523-968e-470e458361f0 service nova] Lock "351b1672-ddb3-4562-9965-d65a7d869e82-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.481508] env[62952]: DEBUG nova.compute.manager [req-ef874528-6b1f-4a08-b2be-809a9889620f req-8bced861-652d-4523-968e-470e458361f0 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] No waiting events found dispatching network-vif-plugged-06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.481997] env[62952]: WARNING nova.compute.manager [req-ef874528-6b1f-4a08-b2be-809a9889620f req-8bced861-652d-4523-968e-470e458361f0 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Received unexpected event network-vif-plugged-06dcb915-bb1f-438c-823b-3fa42d8e6a7d for instance with vm_state building and task_state spawning. [ 889.526062] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6884d72-bc92-4710-85f9-313aaf5e81c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.532576] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd14bf1-8da8-45dc-ad5f-34f3cf082e41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.567679] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b317d7b4-c3f3-42f7-8296-9e38094e0efe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.576336] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8b84a2-8a4e-483c-aa4a-40faae1d98a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.592110] env[62952]: DEBUG nova.compute.provider_tree [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.662646] env[62952]: DEBUG nova.network.neutron [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Successfully updated port: 06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.978847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-16224e22-7091-4381-9f40-3f52f24bb724" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.979195] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 889.979316] env[62952]: DEBUG nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.979485] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.095089] env[62952]: DEBUG nova.scheduler.client.report [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.101163] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.165158] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.165388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquired lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.165481] env[62952]: DEBUG nova.network.neutron [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.601396] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.601841] env[62952]: ERROR nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Traceback (most recent call last): [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self.driver.spawn(context, instance, image_meta, [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] vm_ref = self.build_virtual_machine(instance, [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] vif_infos = vmwarevif.get_vif_info(self._session, [ 890.601841] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] for vif in network_info: [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return self._sync_wrapper(fn, *args, **kwargs) [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self.wait() [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self[:] = self._gt.wait() [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return self._exit_event.wait() [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] result = hub.switch() [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 890.602240] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return self.greenlet.switch() [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] result = function(*args, **kwargs) [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] return func(*args, **kwargs) [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] raise e [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] nwinfo = self.network_api.allocate_for_instance( [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] created_port_ids = self._update_ports_for_instance( [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] with excutils.save_and_reraise_exception(): [ 890.602601] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] self.force_reraise() [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] raise self.value [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] updated_port = self._update_port( [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] _ensure_no_port_binding_failure(port) [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] raise exception.PortBindingFailed(port_id=port['id']) [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] nova.exception.PortBindingFailed: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. [ 890.602960] env[62952]: ERROR nova.compute.manager [instance: b4ea4cff-ada0-4355-a514-d8b179616535] [ 890.603386] env[62952]: DEBUG nova.compute.utils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 890.603719] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.722s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.606645] env[62952]: DEBUG nova.network.neutron [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.607877] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Build of instance b4ea4cff-ada0-4355-a514-d8b179616535 was re-scheduled: Binding failed for port 52e46d51-88f2-4c40-911b-ad3a8f7dc6bb, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 890.608334] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 890.608557] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquiring lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.608706] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Acquired lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.608860] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.676987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "a3c01d4d-9c66-4441-95e8-87998782cc02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.677251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "a3c01d4d-9c66-4441-95e8-87998782cc02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.805491] env[62952]: DEBUG nova.network.neutron [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.991314] env[62952]: DEBUG nova.network.neutron [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Updating instance_info_cache with network_info: [{"id": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "address": "fa:16:3e:f7:b3:f2", "network": {"id": "3539518b-63b4-48b1-90ff-1a0d58fa164a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1315715139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69a5b81b61794ec0a8d884692ec0ed01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06dcb915-bb", "ovs_interfaceid": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.112238] env[62952]: INFO nova.compute.manager [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 16224e22-7091-4381-9f40-3f52f24bb724] Took 1.13 seconds to deallocate network for instance. [ 891.139033] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.259040] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.390133] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7001eab7-9ff0-409c-8427-1a7226433840 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.397509] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06aac444-5036-46d4-ac34-385858a0cc4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.427312] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e9f749-df22-4b13-aacd-eab26f8f030e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.438021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6165ea-e6d9-4395-9833-3c3d8df74851 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.451401] env[62952]: DEBUG nova.compute.provider_tree [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.494939] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Releasing lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.494939] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Instance network_info: |[{"id": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "address": "fa:16:3e:f7:b3:f2", "network": {"id": "3539518b-63b4-48b1-90ff-1a0d58fa164a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1315715139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69a5b81b61794ec0a8d884692ec0ed01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06dcb915-bb", "ovs_interfaceid": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.495181] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:b3:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90878b7b-ddb7-4f47-892b-d6e06f73475f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06dcb915-bb1f-438c-823b-3fa42d8e6a7d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.504210] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Creating folder: Project (69a5b81b61794ec0a8d884692ec0ed01). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.504645] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-293c8c6a-072b-4d08-8d0d-818fb9dcba12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.517153] env[62952]: DEBUG nova.compute.manager [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Received event network-changed-06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.517353] env[62952]: DEBUG nova.compute.manager [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Refreshing instance network info cache due to event network-changed-06dcb915-bb1f-438c-823b-3fa42d8e6a7d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 891.517569] env[62952]: DEBUG oslo_concurrency.lockutils [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] Acquiring lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.517715] env[62952]: DEBUG oslo_concurrency.lockutils [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] Acquired lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.517896] env[62952]: DEBUG nova.network.neutron [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Refreshing network info cache for port 06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.521447] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Created folder: Project (69a5b81b61794ec0a8d884692ec0ed01) in parent group-v290852. [ 891.521625] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Creating folder: Instances. Parent ref: group-v290873. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.522389] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc344651-464a-464d-811b-f7410f2ab689 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.531825] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Created folder: Instances in parent group-v290873. [ 891.532059] env[62952]: DEBUG oslo.service.loopingcall [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.532279] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.532489] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57b19e39-9ab3-4353-89f1-4f65818c0732 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.550779] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.550779] env[62952]: value = "task-1367109" [ 891.550779] env[62952]: _type = "Task" [ 891.550779] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.558819] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367109, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.760557] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Releasing lock "refresh_cache-b4ea4cff-ada0-4355-a514-d8b179616535" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.760735] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 891.760948] env[62952]: DEBUG nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.761185] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 891.780743] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.955022] env[62952]: DEBUG nova.scheduler.client.report [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.062881] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367109, 'name': CreateVM_Task, 'duration_secs': 0.324011} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.064818] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.081019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.081019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.081019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.081019] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a576432-c913-4880-8f8a-5818ca36981c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.087953] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 892.087953] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]527f5048-a977-bc58-f4ac-b0fbcecc85ad" [ 892.087953] env[62952]: _type = "Task" [ 892.087953] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.100051] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527f5048-a977-bc58-f4ac-b0fbcecc85ad, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.100559] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.100932] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.102400] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.102400] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.102400] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.102400] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70980ab5-0151-4fd7-8c27-4743d48566af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.109815] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.110059] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.110780] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-156c31cb-427f-4d29-98a6-5a6c56a61081 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.115860] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 892.115860] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52f3dc8b-3ecd-857e-11ea-c2dfaa2ed934" [ 892.115860] env[62952]: _type = "Task" [ 892.115860] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.127103] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f3dc8b-3ecd-857e-11ea-c2dfaa2ed934, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.167464] env[62952]: INFO nova.scheduler.client.report [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocations for instance 16224e22-7091-4381-9f40-3f52f24bb724 [ 892.286546] env[62952]: DEBUG nova.network.neutron [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.303173] env[62952]: DEBUG nova.network.neutron [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Updated VIF entry in instance network info cache for port 06dcb915-bb1f-438c-823b-3fa42d8e6a7d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.303534] env[62952]: DEBUG nova.network.neutron [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Updating instance_info_cache with network_info: [{"id": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "address": "fa:16:3e:f7:b3:f2", "network": {"id": "3539518b-63b4-48b1-90ff-1a0d58fa164a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1315715139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69a5b81b61794ec0a8d884692ec0ed01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06dcb915-bb", "ovs_interfaceid": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.458669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.855s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.459363] env[62952]: ERROR nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] Traceback (most recent call last): [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self.driver.spawn(context, instance, image_meta, [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self._vmops.spawn(context, instance, image_meta, injected_files, [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] vm_ref = self.build_virtual_machine(instance, [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] vif_infos = vmwarevif.get_vif_info(self._session, [ 892.459363] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] for vif in network_info: [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] return self._sync_wrapper(fn, *args, **kwargs) [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self.wait() [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self[:] = self._gt.wait() [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] return self._exit_event.wait() [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] current.throw(*self._exc) [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 892.460703] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] result = function(*args, **kwargs) [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] return func(*args, **kwargs) [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] raise e [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] nwinfo = self.network_api.allocate_for_instance( [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] created_port_ids = self._update_ports_for_instance( [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] with excutils.save_and_reraise_exception(): [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] self.force_reraise() [ 892.461024] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] raise self.value [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] updated_port = self._update_port( [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] _ensure_no_port_binding_failure(port) [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] raise exception.PortBindingFailed(port_id=port['id']) [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] nova.exception.PortBindingFailed: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. [ 892.461412] env[62952]: ERROR nova.compute.manager [instance: 39948315-6438-48f1-883b-b376e0786650] [ 892.461412] env[62952]: DEBUG nova.compute.utils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 892.461645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.003s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.461645] env[62952]: DEBUG nova.objects.instance [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lazy-loading 'resources' on Instance uuid 34b6c13c-c95b-4cc2-b6eb-cc9374535aec {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.465283] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Build of instance 39948315-6438-48f1-883b-b376e0786650 was re-scheduled: Binding failed for port aa8d3150-71a3-4f36-9bec-823d2e32b95c, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 892.465405] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 892.465578] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquiring lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.466651] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Acquired lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.466651] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.626637] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f3dc8b-3ecd-857e-11ea-c2dfaa2ed934, 'name': SearchDatastore_Task, 'duration_secs': 0.009591} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.627523] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2eadc04-28ee-42bb-bb54-cf760f37975d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.632701] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 892.632701] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52854d96-7bd4-5c41-e6cf-ccfd91168391" [ 892.632701] env[62952]: _type = "Task" [ 892.632701] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.640256] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52854d96-7bd4-5c41-e6cf-ccfd91168391, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.681036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b19bf01e-aa63-4a48-9ebc-c79c8d29dbfc tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "16224e22-7091-4381-9f40-3f52f24bb724" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.238s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.791674] env[62952]: INFO nova.compute.manager [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] [instance: b4ea4cff-ada0-4355-a514-d8b179616535] Took 1.03 seconds to deallocate network for instance. [ 892.806309] env[62952]: DEBUG oslo_concurrency.lockutils [req-7ff47228-73e8-4aae-a0d8-bbee777c88f3 req-cf29ee52-b833-489a-9fdf-50c595ecd088 service nova] Releasing lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.079361] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.151578] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52854d96-7bd4-5c41-e6cf-ccfd91168391, 'name': SearchDatastore_Task, 'duration_secs': 0.02679} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.154170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.154442] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 351b1672-ddb3-4562-9965-d65a7d869e82/351b1672-ddb3-4562-9965-d65a7d869e82.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.155074] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be3d00f6-6354-46b8-9f7e-0da26d354788 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.163439] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 893.163439] env[62952]: value = "task-1367110" [ 893.163439] env[62952]: _type = "Task" [ 893.163439] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.173980] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367110, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.187027] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.190608] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.338127] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c35faf-23db-4c71-9ca6-bc2131eb2aee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.347650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743f82d7-7d01-4f43-a8d0-98de50c499a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.393088] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aaa7881-a775-46ed-a09b-94de1ff8d47a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.404420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8f4cb2-4c68-421d-8996-6fbdb8fb9fe5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.421639] env[62952]: DEBUG nova.compute.provider_tree [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.677628] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367110, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.697102] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Releasing lock "refresh_cache-39948315-6438-48f1-883b-b376e0786650" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.697356] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 893.697538] env[62952]: DEBUG nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.697701] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 893.719862] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.729852] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.823029] env[62952]: INFO nova.scheduler.client.report [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Deleted allocations for instance b4ea4cff-ada0-4355-a514-d8b179616535 [ 893.925669] env[62952]: DEBUG nova.scheduler.client.report [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.178866] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367110, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547923} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.179130] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 351b1672-ddb3-4562-9965-d65a7d869e82/351b1672-ddb3-4562-9965-d65a7d869e82.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.179365] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.179598] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aeef417f-5f57-4bd0-b3f6-82e033a4b88b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.187377] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 894.187377] env[62952]: value = "task-1367111" [ 894.187377] env[62952]: _type = "Task" [ 894.187377] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.197552] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.207789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.208049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.225191] env[62952]: DEBUG nova.network.neutron [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.333933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f6a0913f-7a03-4cc6-bf66-0886aa3d8007 tempest-AttachInterfacesV270Test-1651024537 tempest-AttachInterfacesV270Test-1651024537-project-member] Lock "b4ea4cff-ada0-4355-a514-d8b179616535" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.983s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.436942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.439301] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.335s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.440748] env[62952]: INFO nova.compute.claims [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.457729] env[62952]: INFO nova.scheduler.client.report [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Deleted allocations for instance 34b6c13c-c95b-4cc2-b6eb-cc9374535aec [ 894.697263] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.730471] env[62952]: INFO nova.compute.manager [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] [instance: 39948315-6438-48f1-883b-b376e0786650] Took 1.03 seconds to deallocate network for instance. [ 894.829929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.830231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.837043] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.965741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-efb494a1-cf51-4ccd-bb1a-317c1732f135 tempest-ServerDiagnosticsV248Test-1612581296 tempest-ServerDiagnosticsV248Test-1612581296-project-member] Lock "34b6c13c-c95b-4cc2-b6eb-cc9374535aec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.202s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.197640] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.585558} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.198111] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.198701] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1647a92c-e5d9-490e-bda1-567c5559da00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.222023] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 351b1672-ddb3-4562-9965-d65a7d869e82/351b1672-ddb3-4562-9965-d65a7d869e82.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.222790] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08b8a8e8-e2ba-421e-b4dd-c571e3620970 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.246331] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 895.246331] env[62952]: value = "task-1367112" [ 895.246331] env[62952]: _type = "Task" [ 895.246331] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.255953] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367112, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.368194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.732503] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad7aa69-1f78-446e-a574-406ca43a9681 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.741118] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e405d8ff-bfe5-49e1-8e07-a2d47eb0d928 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.796176] env[62952]: INFO nova.scheduler.client.report [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Deleted allocations for instance 39948315-6438-48f1-883b-b376e0786650 [ 895.804169] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d83284f-b512-4ff4-8688-6ee346b08c42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.816146] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367112, 'name': ReconfigVM_Task, 'duration_secs': 0.282968} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.816146] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 351b1672-ddb3-4562-9965-d65a7d869e82/351b1672-ddb3-4562-9965-d65a7d869e82.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.816787] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3455f90-5ab4-42a6-a308-c6672586ce35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.823998] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d4b495-c12c-4526-92c9-1cd4da6ec1ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.829398] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 895.829398] env[62952]: value = "task-1367113" [ 895.829398] env[62952]: _type = "Task" [ 895.829398] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.844317] env[62952]: DEBUG nova.compute.provider_tree [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.851414] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367113, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.320145] env[62952]: DEBUG oslo_concurrency.lockutils [None req-386ac631-eefe-4c3d-a720-cee3ecb921e4 tempest-SecurityGroupsTestJSON-1309146595 tempest-SecurityGroupsTestJSON-1309146595-project-member] Lock "39948315-6438-48f1-883b-b376e0786650" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.093s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.340379] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367113, 'name': Rename_Task, 'duration_secs': 0.161216} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.340714] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.340951] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-196ba173-07f7-4bbc-acb0-f3d93b162e27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.352916] env[62952]: DEBUG nova.scheduler.client.report [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.359055] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 896.359055] env[62952]: value = "task-1367114" [ 896.359055] env[62952]: _type = "Task" [ 896.359055] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.370482] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367114, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.824286] env[62952]: DEBUG nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 896.861638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.862306] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.865645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.962s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.867141] env[62952]: INFO nova.compute.claims [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.886082] env[62952]: DEBUG oslo_vmware.api [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367114, 'name': PowerOnVM_Task, 'duration_secs': 0.498385} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.886584] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.886786] env[62952]: INFO nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Took 8.05 seconds to spawn the instance on the hypervisor. [ 896.886958] env[62952]: DEBUG nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.887893] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349bc56f-956e-4aa8-b8a1-e088f19802de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.352068] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.380449] env[62952]: DEBUG nova.compute.utils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.381718] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.381878] env[62952]: DEBUG nova.network.neutron [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.405854] env[62952]: INFO nova.compute.manager [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Took 38.71 seconds to build instance. [ 897.441970] env[62952]: DEBUG nova.policy [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca180ffa174d4c15b9d0afa723a66149', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d8eb11ac5ba4149942207c00ecc2652', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.888738] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.908716] env[62952]: DEBUG oslo_concurrency.lockutils [None req-24be8fc7-f9af-4fee-8c99-2e7dbc9af342 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.663s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.933840] env[62952]: DEBUG nova.network.neutron [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Successfully created port: e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.097538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "351b1672-ddb3-4562-9965-d65a7d869e82" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.097755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.097978] env[62952]: INFO nova.compute.manager [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Rebooting instance [ 898.188868] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9cefb68-4406-4c8c-8bd5-8d6ec99222be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.200129] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67eee75-29de-49c6-b528-775c44f07c4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.229467] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4bdd97-c0d1-45c1-ad9f-69b467b84c67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.237122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d866d1f-5ea3-45b1-819f-ae84ae1572e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.251580] env[62952]: DEBUG nova.compute.provider_tree [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.413469] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.623106] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.623297] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquired lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.623471] env[62952]: DEBUG nova.network.neutron [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.756147] env[62952]: DEBUG nova.scheduler.client.report [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.901665] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.935812] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.936103] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.936267] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.936453] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.936598] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.936750] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.937009] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.937427] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.937427] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.937579] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.937778] env[62952]: DEBUG nova.virt.hardware [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.938630] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2347013-aa2f-4483-ab4d-18aab12bf626 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.947088] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.948347] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1839b0e-d8cf-475d-95f6-634cbbcd5d82 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.261049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.262260] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.264646] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.183s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.312882] env[62952]: DEBUG nova.network.neutron [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Updating instance_info_cache with network_info: [{"id": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "address": "fa:16:3e:f7:b3:f2", "network": {"id": "3539518b-63b4-48b1-90ff-1a0d58fa164a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1315715139-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69a5b81b61794ec0a8d884692ec0ed01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90878b7b-ddb7-4f47-892b-d6e06f73475f", "external-id": "nsx-vlan-transportzone-849", "segmentation_id": 849, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06dcb915-bb", "ovs_interfaceid": "06dcb915-bb1f-438c-823b-3fa42d8e6a7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.728235] env[62952]: DEBUG nova.network.neutron [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Successfully updated port: e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.768706] env[62952]: DEBUG nova.compute.utils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.773943] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.774833] env[62952]: DEBUG nova.network.neutron [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.795248] env[62952]: DEBUG nova.compute.manager [req-ce5afac0-c5ec-46b7-918a-ee8b98103e47 req-4664c193-74de-4fe2-a610-e11bd237f707 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Received event network-vif-plugged-e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.795248] env[62952]: DEBUG oslo_concurrency.lockutils [req-ce5afac0-c5ec-46b7-918a-ee8b98103e47 req-4664c193-74de-4fe2-a610-e11bd237f707 service nova] Acquiring lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.795248] env[62952]: DEBUG oslo_concurrency.lockutils [req-ce5afac0-c5ec-46b7-918a-ee8b98103e47 req-4664c193-74de-4fe2-a610-e11bd237f707 service nova] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.795248] env[62952]: DEBUG oslo_concurrency.lockutils [req-ce5afac0-c5ec-46b7-918a-ee8b98103e47 req-4664c193-74de-4fe2-a610-e11bd237f707 service nova] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.795248] env[62952]: DEBUG nova.compute.manager [req-ce5afac0-c5ec-46b7-918a-ee8b98103e47 req-4664c193-74de-4fe2-a610-e11bd237f707 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] No waiting events found dispatching network-vif-plugged-e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.795499] env[62952]: WARNING nova.compute.manager [req-ce5afac0-c5ec-46b7-918a-ee8b98103e47 req-4664c193-74de-4fe2-a610-e11bd237f707 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Received unexpected event network-vif-plugged-e99d8558-48fe-47e0-9534-c7953ad6fa8c for instance with vm_state building and task_state spawning. [ 899.815504] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Releasing lock "refresh_cache-351b1672-ddb3-4562-9965-d65a7d869e82" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.819578] env[62952]: DEBUG nova.compute.manager [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.820753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df222230-ae2a-4cbd-a7bb-3207739dae6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.848033] env[62952]: DEBUG nova.policy [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0194e5d37e4945ad7f2436032d2b60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6d1dbb357954f40995c72695f446232', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.067302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24d8b8e-a2fc-4903-bb23-2b5d4e3f0dc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.075666] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6469e019-27ea-4bfd-b093-5f1c2a41fb5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.111413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6a48dd-9b1d-456b-adfa-759085c5197f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.121176] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c26736-f3d5-4263-a42e-8314514a25d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.139936] env[62952]: DEBUG nova.compute.provider_tree [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.167148] env[62952]: DEBUG nova.network.neutron [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Successfully created port: cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.231338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.231487] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquired lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.231702] env[62952]: DEBUG nova.network.neutron [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.278021] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.644238] env[62952]: DEBUG nova.scheduler.client.report [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.771817] env[62952]: DEBUG nova.network.neutron [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.843744] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1320a3ed-97c9-40cc-ab92-b5dada425fd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.853822] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Doing hard reboot of VM {{(pid=62952) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 900.853952] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-57ad713f-6b11-4531-a94e-69ee1aae8d04 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.861585] env[62952]: DEBUG oslo_vmware.api [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 900.861585] env[62952]: value = "task-1367115" [ 900.861585] env[62952]: _type = "Task" [ 900.861585] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.869098] env[62952]: DEBUG oslo_vmware.api [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367115, 'name': ResetVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.943113] env[62952]: DEBUG nova.network.neutron [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updating instance_info_cache with network_info: [{"id": "e99d8558-48fe-47e0-9534-c7953ad6fa8c", "address": "fa:16:3e:1f:25:a6", "network": {"id": "9f808e18-a9e9-4199-8176-cdfb17cbee40", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1970219864-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d8eb11ac5ba4149942207c00ecc2652", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape99d8558-48", "ovs_interfaceid": "e99d8558-48fe-47e0-9534-c7953ad6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.149073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.149760] env[62952]: ERROR nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Traceback (most recent call last): [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self.driver.spawn(context, instance, image_meta, [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self._vmops.spawn(context, instance, image_meta, injected_files, [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] vm_ref = self.build_virtual_machine(instance, [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] vif_infos = vmwarevif.get_vif_info(self._session, [ 901.149760] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] for vif in network_info: [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return self._sync_wrapper(fn, *args, **kwargs) [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self.wait() [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self[:] = self._gt.wait() [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return self._exit_event.wait() [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] result = hub.switch() [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 901.150101] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return self.greenlet.switch() [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] result = function(*args, **kwargs) [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] return func(*args, **kwargs) [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] raise e [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] nwinfo = self.network_api.allocate_for_instance( [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] created_port_ids = self._update_ports_for_instance( [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] with excutils.save_and_reraise_exception(): [ 901.150402] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] self.force_reraise() [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] raise self.value [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] updated_port = self._update_port( [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] _ensure_no_port_binding_failure(port) [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] raise exception.PortBindingFailed(port_id=port['id']) [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] nova.exception.PortBindingFailed: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. [ 901.150699] env[62952]: ERROR nova.compute.manager [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] [ 901.150959] env[62952]: DEBUG nova.compute.utils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 901.152148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.676s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.155546] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Build of instance 69c93772-822a-4f5b-b2d1-95e82a46f286 was re-scheduled: Binding failed for port 3951bf5b-4bf7-4759-acac-0c83678ccf69, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 901.156073] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 901.156358] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquiring lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.156571] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Acquired lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.156784] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.283712] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.311479] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.311913] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.311913] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.312119] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.312266] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.312412] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.312616] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.312841] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.313063] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.313267] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.313453] env[62952]: DEBUG nova.virt.hardware [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.314334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a15842-51e7-4b11-a617-a6a01b5d4f81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.323725] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9ca347-256c-46b9-beac-d79b0ad25178 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.370430] env[62952]: DEBUG oslo_vmware.api [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367115, 'name': ResetVM_Task, 'duration_secs': 0.097395} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.370693] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Did hard reboot of VM {{(pid=62952) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 901.370864] env[62952]: DEBUG nova.compute.manager [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.371609] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fe824a-e031-4ca3-9642-54dd0c521b33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.446275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Releasing lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.446658] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Instance network_info: |[{"id": "e99d8558-48fe-47e0-9534-c7953ad6fa8c", "address": "fa:16:3e:1f:25:a6", "network": {"id": "9f808e18-a9e9-4199-8176-cdfb17cbee40", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1970219864-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d8eb11ac5ba4149942207c00ecc2652", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape99d8558-48", "ovs_interfaceid": "e99d8558-48fe-47e0-9534-c7953ad6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.446967] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:25:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06cc7c49-c46c-4c1e-bf51-77e9ea802c40', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e99d8558-48fe-47e0-9534-c7953ad6fa8c', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.456680] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Creating folder: Project (2d8eb11ac5ba4149942207c00ecc2652). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.456976] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a2c686e-e4e7-44ba-858f-a8936f081f58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.469039] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Created folder: Project (2d8eb11ac5ba4149942207c00ecc2652) in parent group-v290852. [ 901.469492] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Creating folder: Instances. Parent ref: group-v290876. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.469492] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6990c489-460a-4c7e-87ce-c46cfe293fde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.479106] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Created folder: Instances in parent group-v290876. [ 901.479349] env[62952]: DEBUG oslo.service.loopingcall [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.479544] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.479752] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a85fc300-df1d-4710-b936-913402e8b31d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.500034] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.500034] env[62952]: value = "task-1367118" [ 901.500034] env[62952]: _type = "Task" [ 901.500034] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.508987] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367118, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.679051] env[62952]: DEBUG nova.network.neutron [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Successfully updated port: cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.685250] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 901.791758] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.830837] env[62952]: DEBUG nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Received event network-changed-e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.831487] env[62952]: DEBUG nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Refreshing instance network info cache due to event network-changed-e99d8558-48fe-47e0-9534-c7953ad6fa8c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.831841] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Acquiring lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.832019] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Acquired lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.832192] env[62952]: DEBUG nova.network.neutron [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Refreshing network info cache for port e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.885708] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee4d4701-c6b0-4f6f-a276-f7341748605c tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.787s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.954026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b0fd80-edcd-4efe-9893-7c05cddbbaab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.963202] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a362805-08df-4153-a1bb-159838a5de3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.995898] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9daeee-0516-44bd-afb6-c4db508e201e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.007577] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e7cdd2-5d86-4e96-a36d-6bb98046800a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.022634] env[62952]: DEBUG nova.compute.provider_tree [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.030254] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367118, 'name': CreateVM_Task, 'duration_secs': 0.305708} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.030555] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.031274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.031545] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.031746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.032021] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eea11bf8-552a-4e98-a022-f2a1290ca2e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.041019] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 902.041019] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52207db2-209b-e1b7-c665-e494f978bd75" [ 902.041019] env[62952]: _type = "Task" [ 902.041019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.045868] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52207db2-209b-e1b7-c665-e494f978bd75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.183537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.184842] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.185099] env[62952]: DEBUG nova.network.neutron [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.295062] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Releasing lock "refresh_cache-69c93772-822a-4f5b-b2d1-95e82a46f286" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.295428] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 902.295667] env[62952]: DEBUG nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.295845] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 902.319736] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.533159] env[62952]: DEBUG nova.scheduler.client.report [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.549054] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52207db2-209b-e1b7-c665-e494f978bd75, 'name': SearchDatastore_Task, 'duration_secs': 0.012768} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.549054] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.549721] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.549810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.549912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.550138] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.550439] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98a63882-1a23-4ea1-84e3-5ddb1f7da0d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.559481] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.559557] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.560319] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-966bea41-404d-4857-b81d-a191619fa164 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.569865] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 902.569865] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ab82b0-7115-a1b6-0f1d-9d10377411d7" [ 902.569865] env[62952]: _type = "Task" [ 902.569865] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.578867] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ab82b0-7115-a1b6-0f1d-9d10377411d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.613382] env[62952]: DEBUG nova.network.neutron [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updated VIF entry in instance network info cache for port e99d8558-48fe-47e0-9534-c7953ad6fa8c. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.613382] env[62952]: DEBUG nova.network.neutron [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updating instance_info_cache with network_info: [{"id": "e99d8558-48fe-47e0-9534-c7953ad6fa8c", "address": "fa:16:3e:1f:25:a6", "network": {"id": "9f808e18-a9e9-4199-8176-cdfb17cbee40", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1970219864-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2d8eb11ac5ba4149942207c00ecc2652", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape99d8558-48", "ovs_interfaceid": "e99d8558-48fe-47e0-9534-c7953ad6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.725191] env[62952]: DEBUG nova.network.neutron [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.823337] env[62952]: DEBUG nova.network.neutron [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.962094] env[62952]: DEBUG nova.network.neutron [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Updating instance_info_cache with network_info: [{"id": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "address": "fa:16:3e:07:66:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea71c6a-30", "ovs_interfaceid": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.988291] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "351b1672-ddb3-4562-9965-d65a7d869e82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.988560] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.988772] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "351b1672-ddb3-4562-9965-d65a7d869e82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.988953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.989134] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.993248] env[62952]: INFO nova.compute.manager [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Terminating instance [ 902.996293] env[62952]: DEBUG nova.compute.manager [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.996293] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.996948] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8c661d-591e-4155-84db-ece1603feaf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.007194] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.007194] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e01a855-bbfe-4c7d-b350-f9a65a705cc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.014463] env[62952]: DEBUG oslo_vmware.api [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 903.014463] env[62952]: value = "task-1367119" [ 903.014463] env[62952]: _type = "Task" [ 903.014463] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.024469] env[62952]: DEBUG oslo_vmware.api [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367119, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.037822] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.885s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.040259] env[62952]: ERROR nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Traceback (most recent call last): [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self.driver.spawn(context, instance, image_meta, [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self._vmops.spawn(context, instance, image_meta, injected_files, [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] vm_ref = self.build_virtual_machine(instance, [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] vif_infos = vmwarevif.get_vif_info(self._session, [ 903.040259] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] for vif in network_info: [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] return self._sync_wrapper(fn, *args, **kwargs) [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self.wait() [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self[:] = self._gt.wait() [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] return self._exit_event.wait() [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] current.throw(*self._exc) [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 903.040532] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] result = function(*args, **kwargs) [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] return func(*args, **kwargs) [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] raise e [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] nwinfo = self.network_api.allocate_for_instance( [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] created_port_ids = self._update_ports_for_instance( [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] with excutils.save_and_reraise_exception(): [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] self.force_reraise() [ 903.040833] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] raise self.value [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] updated_port = self._update_port( [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] _ensure_no_port_binding_failure(port) [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] raise exception.PortBindingFailed(port_id=port['id']) [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] nova.exception.PortBindingFailed: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. [ 903.041210] env[62952]: ERROR nova.compute.manager [instance: 62d47ebb-8782-4062-a27e-556c758ed475] [ 903.041210] env[62952]: DEBUG nova.compute.utils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 903.046022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.226s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.047081] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Build of instance 62d47ebb-8782-4062-a27e-556c758ed475 was re-scheduled: Binding failed for port fe3a90cb-d12d-4ef6-a2fa-c058530e56a3, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 903.047648] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 903.048444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.048444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.048577] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.087017] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ab82b0-7115-a1b6-0f1d-9d10377411d7, 'name': SearchDatastore_Task, 'duration_secs': 0.0087} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.087875] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c89553d4-df47-4355-8c95-1224a35f358f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.093440] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 903.093440] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52f04ab9-075c-c782-ca52-1fc9e8146608" [ 903.093440] env[62952]: _type = "Task" [ 903.093440] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.102746] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f04ab9-075c-c782-ca52-1fc9e8146608, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.118093] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Releasing lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.118362] env[62952]: DEBUG nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Received event network-vif-plugged-cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.118550] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Acquiring lock "a63ec2ac-4484-4360-962b-105a5338c1b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.118744] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.118916] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.119100] env[62952]: DEBUG nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] No waiting events found dispatching network-vif-plugged-cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.119264] env[62952]: WARNING nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Received unexpected event network-vif-plugged-cea71c6a-30a0-4f56-bc79-6b103c4b59c3 for instance with vm_state building and task_state spawning. [ 903.119420] env[62952]: DEBUG nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Received event network-changed-cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.119569] env[62952]: DEBUG nova.compute.manager [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Refreshing instance network info cache due to event network-changed-cea71c6a-30a0-4f56-bc79-6b103c4b59c3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.119724] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Acquiring lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.327429] env[62952]: INFO nova.compute.manager [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] [instance: 69c93772-822a-4f5b-b2d1-95e82a46f286] Took 1.03 seconds to deallocate network for instance. [ 903.465188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.465961] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Instance network_info: |[{"id": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "address": "fa:16:3e:07:66:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea71c6a-30", "ovs_interfaceid": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.465961] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Acquired lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.466162] env[62952]: DEBUG nova.network.neutron [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Refreshing network info cache for port cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.467186] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:66:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c0d5204b-f60e-4830-84c8-2fe246c28202', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cea71c6a-30a0-4f56-bc79-6b103c4b59c3', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.475078] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Creating folder: Project (f6d1dbb357954f40995c72695f446232). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.479023] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8af76c12-3772-4882-ac59-030402f9fb92 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.491396] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Created folder: Project (f6d1dbb357954f40995c72695f446232) in parent group-v290852. [ 903.491594] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Creating folder: Instances. Parent ref: group-v290879. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.491842] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57df4454-bf88-42f0-bf36-bf6f2b222e7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.501562] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Created folder: Instances in parent group-v290879. [ 903.501718] env[62952]: DEBUG oslo.service.loopingcall [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.501945] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.502177] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43848143-1f7a-4672-a1fd-577e10f5be85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.530070] env[62952]: DEBUG oslo_vmware.api [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367119, 'name': PowerOffVM_Task, 'duration_secs': 0.223252} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.531405] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.531699] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.531988] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.531988] env[62952]: value = "task-1367122" [ 903.531988] env[62952]: _type = "Task" [ 903.531988] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.532290] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7f878bd-2de1-4791-adb1-2e1084998bd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.577487] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.601487] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.601985] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.601985] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Deleting the datastore file [datastore2] 351b1672-ddb3-4562-9965-d65a7d869e82 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.602867] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d916ab8a-5d63-4d67-9283-6a0c92b91155 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.608363] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f04ab9-075c-c782-ca52-1fc9e8146608, 'name': SearchDatastore_Task, 'duration_secs': 0.009949} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.611422] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.611800] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] d2faf1e0-cfd1-4d87-ba77-0af92dc16643/d2faf1e0-cfd1-4d87-ba77-0af92dc16643.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.614570] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38261d3e-d6e3-4d7e-a8a3-5bc24d003c4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.620039] env[62952]: DEBUG oslo_vmware.api [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for the task: (returnval){ [ 903.620039] env[62952]: value = "task-1367124" [ 903.620039] env[62952]: _type = "Task" [ 903.620039] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.628217] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 903.628217] env[62952]: value = "task-1367125" [ 903.628217] env[62952]: _type = "Task" [ 903.628217] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.633637] env[62952]: DEBUG oslo_vmware.api [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.640697] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.767483] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.850354] env[62952]: DEBUG nova.network.neutron [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Updated VIF entry in instance network info cache for port cea71c6a-30a0-4f56-bc79-6b103c4b59c3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.850354] env[62952]: DEBUG nova.network.neutron [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Updating instance_info_cache with network_info: [{"id": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "address": "fa:16:3e:07:66:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea71c6a-30", "ovs_interfaceid": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.911562] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2fd96b-77f3-464b-afac-c0d7d5e91791 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.924022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3e74dc-4c3e-4732-86b8-2c79a7e5c806 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.957727] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1722314-17e6-427d-bb1f-0f62ecc69ca4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.966974] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13f6f2c-4eb2-406b-ad08-b4f5da3bef53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.987915] env[62952]: DEBUG nova.compute.provider_tree [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.043862] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367122, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.136245] env[62952]: DEBUG oslo_vmware.api [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Task: {'id': task-1367124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150572} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.137045] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.137124] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.137288] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.137474] env[62952]: INFO nova.compute.manager [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Took 1.14 seconds to destroy the instance on the hypervisor. [ 904.137695] env[62952]: DEBUG oslo.service.loopingcall [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.137894] env[62952]: DEBUG nova.compute.manager [-] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.138055] env[62952]: DEBUG nova.network.neutron [-] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.142570] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367125, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48168} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.143150] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] d2faf1e0-cfd1-4d87-ba77-0af92dc16643/d2faf1e0-cfd1-4d87-ba77-0af92dc16643.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.143320] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.143585] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78ab1b57-6e49-4f99-be72-90e71e0ec0e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.154442] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 904.154442] env[62952]: value = "task-1367126" [ 904.154442] env[62952]: _type = "Task" [ 904.154442] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.163027] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367126, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.274272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-62d47ebb-8782-4062-a27e-556c758ed475" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.274607] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 904.274815] env[62952]: DEBUG nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.275000] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.299707] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 904.349448] env[62952]: DEBUG oslo_concurrency.lockutils [req-65ecf5e0-70b8-44ba-a3a3-6398dc8ea01c req-4e06fbae-da30-4629-882f-98ae7670b256 service nova] Releasing lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.365919] env[62952]: INFO nova.scheduler.client.report [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Deleted allocations for instance 69c93772-822a-4f5b-b2d1-95e82a46f286 [ 904.505050] env[62952]: DEBUG nova.compute.manager [req-c84cae4f-ce5b-4cb9-93b2-296e281ae008 req-4858ed61-3d4f-46f1-a400-f20942224355 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Received event network-vif-deleted-06dcb915-bb1f-438c-823b-3fa42d8e6a7d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.505050] env[62952]: INFO nova.compute.manager [req-c84cae4f-ce5b-4cb9-93b2-296e281ae008 req-4858ed61-3d4f-46f1-a400-f20942224355 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Neutron deleted interface 06dcb915-bb1f-438c-823b-3fa42d8e6a7d; detaching it from the instance and deleting it from the info cache [ 904.505050] env[62952]: DEBUG nova.network.neutron [req-c84cae4f-ce5b-4cb9-93b2-296e281ae008 req-4858ed61-3d4f-46f1-a400-f20942224355 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.506271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.506479] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.513365] env[62952]: ERROR nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [req-73edf33f-1700-467e-9086-ce28096fd122] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-73edf33f-1700-467e-9086-ce28096fd122"}]}: nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 904.539628] env[62952]: DEBUG nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 904.548651] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367122, 'name': CreateVM_Task, 'duration_secs': 0.514785} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.548651] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.549204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.549440] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.550278] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.550791] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aec0a43-b1a1-4bed-965b-1e1ee21c1bd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.556527] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 904.556527] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52461fab-8cc1-8869-ed9d-87bf0e4a3416" [ 904.556527] env[62952]: _type = "Task" [ 904.556527] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.557620] env[62952]: DEBUG nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 904.557912] env[62952]: DEBUG nova.compute.provider_tree [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.568656] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52461fab-8cc1-8869-ed9d-87bf0e4a3416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.570470] env[62952]: DEBUG nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 904.591874] env[62952]: DEBUG nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 904.661012] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367126, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071688} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.661361] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.662208] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bf8ddf-4dc2-4b95-a595-d5b292080ab0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.684370] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] d2faf1e0-cfd1-4d87-ba77-0af92dc16643/d2faf1e0-cfd1-4d87-ba77-0af92dc16643.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.687150] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9f26a9f-d024-4949-bbb4-383868cb7000 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.709398] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 904.709398] env[62952]: value = "task-1367127" [ 904.709398] env[62952]: _type = "Task" [ 904.709398] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.719721] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367127, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.802868] env[62952]: DEBUG nova.network.neutron [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.873477] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1c0c7ede-00d0-4f08-ad46-f019e7259470 tempest-ServersNegativeTestJSON-1303968808 tempest-ServersNegativeTestJSON-1303968808-project-member] Lock "69c93772-822a-4f5b-b2d1-95e82a46f286" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.590s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.882987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6cae06-0936-4269-8403-ac393d26e279 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.893159] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfc8a7d-fbd4-4479-bf6f-6672b18232f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.924746] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a377b931-04fb-4edc-a52b-c4803a90f727 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.933125] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0d0696-778f-4050-b18a-2e8376bef384 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.945458] env[62952]: DEBUG nova.compute.provider_tree [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.973838] env[62952]: DEBUG nova.network.neutron [-] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.007758] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bab13186-7265-4943-885f-1b24a38dfc35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.017492] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8744951b-7081-4646-b590-50cecf93940e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.041417] env[62952]: DEBUG nova.compute.manager [req-c84cae4f-ce5b-4cb9-93b2-296e281ae008 req-4858ed61-3d4f-46f1-a400-f20942224355 service nova] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Detach interface failed, port_id=06dcb915-bb1f-438c-823b-3fa42d8e6a7d, reason: Instance 351b1672-ddb3-4562-9965-d65a7d869e82 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 905.069930] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52461fab-8cc1-8869-ed9d-87bf0e4a3416, 'name': SearchDatastore_Task, 'duration_secs': 0.031968} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.070234] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.070412] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.070640] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.070784] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.070961] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.071246] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aef9059b-be62-4c39-89ab-d7715341fefc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.078132] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.078312] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.078961] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b42f8d6-7c2a-475a-872b-a59226a11673 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.083484] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 905.083484] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52403082-463b-2200-5391-87ded6939752" [ 905.083484] env[62952]: _type = "Task" [ 905.083484] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.092537] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52403082-463b-2200-5391-87ded6939752, 'name': SearchDatastore_Task, 'duration_secs': 0.006898} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.093226] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3587c8b2-0603-4bca-90b8-bf3cb0886bc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.097573] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 905.097573] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c01580-fb54-7b85-a4dc-1b8816bba7d6" [ 905.097573] env[62952]: _type = "Task" [ 905.097573] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.104444] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c01580-fb54-7b85-a4dc-1b8816bba7d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.219286] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367127, 'name': ReconfigVM_Task, 'duration_secs': 0.340568} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.219561] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Reconfigured VM instance instance-0000003b to attach disk [datastore1] d2faf1e0-cfd1-4d87-ba77-0af92dc16643/d2faf1e0-cfd1-4d87-ba77-0af92dc16643.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.220203] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28a5df79-1178-4625-a814-f10b57ed50c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.226670] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 905.226670] env[62952]: value = "task-1367128" [ 905.226670] env[62952]: _type = "Task" [ 905.226670] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.237025] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367128, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.306025] env[62952]: INFO nova.compute.manager [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 62d47ebb-8782-4062-a27e-556c758ed475] Took 1.03 seconds to deallocate network for instance. [ 905.376195] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.476688] env[62952]: INFO nova.compute.manager [-] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Took 1.34 seconds to deallocate network for instance. [ 905.482564] env[62952]: DEBUG nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 90 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 905.482814] env[62952]: DEBUG nova.compute.provider_tree [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 90 to 91 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 905.483031] env[62952]: DEBUG nova.compute.provider_tree [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.608381] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c01580-fb54-7b85-a4dc-1b8816bba7d6, 'name': SearchDatastore_Task, 'duration_secs': 0.007326} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.608629] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.608880] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] a63ec2ac-4484-4360-962b-105a5338c1b8/a63ec2ac-4484-4360-962b-105a5338c1b8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.609151] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6306048-bef0-4e21-a4c1-02fccc62897f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.616483] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 905.616483] env[62952]: value = "task-1367129" [ 905.616483] env[62952]: _type = "Task" [ 905.616483] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.625248] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.736547] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367128, 'name': Rename_Task, 'duration_secs': 0.134725} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.736547] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.736547] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b0e4e52-c96a-4c05-bd9d-ba8021b260e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.742733] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 905.742733] env[62952]: value = "task-1367130" [ 905.742733] env[62952]: _type = "Task" [ 905.742733] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.754804] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367130, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.900783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.987595] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.945s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.988677] env[62952]: ERROR nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Traceback (most recent call last): [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self.driver.spawn(context, instance, image_meta, [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] vm_ref = self.build_virtual_machine(instance, [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] vif_infos = vmwarevif.get_vif_info(self._session, [ 905.988677] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] for vif in network_info: [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return self._sync_wrapper(fn, *args, **kwargs) [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self.wait() [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self[:] = self._gt.wait() [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return self._exit_event.wait() [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] result = hub.switch() [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 905.989094] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return self.greenlet.switch() [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] result = function(*args, **kwargs) [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] return func(*args, **kwargs) [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] raise e [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] nwinfo = self.network_api.allocate_for_instance( [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] created_port_ids = self._update_ports_for_instance( [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] with excutils.save_and_reraise_exception(): [ 905.989438] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] self.force_reraise() [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] raise self.value [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] updated_port = self._update_port( [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] _ensure_no_port_binding_failure(port) [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] raise exception.PortBindingFailed(port_id=port['id']) [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] nova.exception.PortBindingFailed: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. [ 905.989811] env[62952]: ERROR nova.compute.manager [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] [ 905.990153] env[62952]: DEBUG nova.compute.utils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. {{(pid=62952) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 905.990983] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.991607] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Build of instance 56aa8186-2bcb-47fc-a832-31654cbf218b was re-scheduled: Binding failed for port 1f6c79c1-d3bd-4eb2-971c-0e6da8192f65, please check neutron logs for more information. {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 905.992130] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Unplugging VIFs for instance {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 905.992386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquiring lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.992536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Acquired lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.992704] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.997807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.761s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.999425] env[62952]: INFO nova.compute.claims [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.131652] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430692} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.131652] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] a63ec2ac-4484-4360-962b-105a5338c1b8/a63ec2ac-4484-4360-962b-105a5338c1b8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.131652] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.131652] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45de27b3-821a-47f2-b073-b15360ed0ccd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.136935] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 906.136935] env[62952]: value = "task-1367131" [ 906.136935] env[62952]: _type = "Task" [ 906.136935] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.144477] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367131, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.252804] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367130, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.344070] env[62952]: INFO nova.scheduler.client.report [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleted allocations for instance 62d47ebb-8782-4062-a27e-556c758ed475 [ 906.528500] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.632190] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.647672] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367131, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067278} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.647931] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.648710] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8329ee2a-fe4c-4ad0-aed8-ad98da31a279 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.670613] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] a63ec2ac-4484-4360-962b-105a5338c1b8/a63ec2ac-4484-4360-962b-105a5338c1b8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.670828] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fe750d4-2f11-421a-88e4-6b9606006ba4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.693256] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 906.693256] env[62952]: value = "task-1367132" [ 906.693256] env[62952]: _type = "Task" [ 906.693256] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.702721] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367132, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.753899] env[62952]: DEBUG oslo_vmware.api [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367130, 'name': PowerOnVM_Task, 'duration_secs': 0.62419} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.754742] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.754742] env[62952]: INFO nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Took 7.85 seconds to spawn the instance on the hypervisor. [ 906.754742] env[62952]: DEBUG nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.755306] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8804af5c-8b3e-404f-8a81-daa550c92d1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.857806] env[62952]: DEBUG oslo_concurrency.lockutils [None req-eec5019f-db6b-4609-910e-e44222b4b9b6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "62d47ebb-8782-4062-a27e-556c758ed475" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.989s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.134901] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Releasing lock "refresh_cache-56aa8186-2bcb-47fc-a832-31654cbf218b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.135206] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62952) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 907.135381] env[62952]: DEBUG nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.135556] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.166204] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.206077] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367132, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.239091] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e86fbc-bbbe-41ec-9b86-85f4f76acf88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.246895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62700ce2-7891-413f-9adf-ea757676cae8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.282866] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028446ed-3c94-4817-96ed-bd48dba8c977 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.287870] env[62952]: INFO nova.compute.manager [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Took 29.20 seconds to build instance. [ 907.292183] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51370514-29d7-49bb-8148-4915856b090a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.306183] env[62952]: DEBUG nova.compute.provider_tree [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.360268] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 907.669366] env[62952]: DEBUG nova.network.neutron [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.705704] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367132, 'name': ReconfigVM_Task, 'duration_secs': 0.986806} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.705704] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Reconfigured VM instance instance-0000003c to attach disk [datastore1] a63ec2ac-4484-4360-962b-105a5338c1b8/a63ec2ac-4484-4360-962b-105a5338c1b8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.705704] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d96c3f36-deef-40a0-9ab1-e1865df3afbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.712967] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 907.712967] env[62952]: value = "task-1367133" [ 907.712967] env[62952]: _type = "Task" [ 907.712967] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.721442] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367133, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.791031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a1222f8a-2986-4ee4-bc18-821f7a1c4b12 tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.782s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.811039] env[62952]: DEBUG nova.scheduler.client.report [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.883527] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.172134] env[62952]: INFO nova.compute.manager [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] [instance: 56aa8186-2bcb-47fc-a832-31654cbf218b] Took 1.04 seconds to deallocate network for instance. [ 908.226791] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367133, 'name': Rename_Task, 'duration_secs': 0.155313} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.226791] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.226791] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cedf7518-eb78-4fa8-8268-c88cdc6cfd41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.231587] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 908.231587] env[62952]: value = "task-1367134" [ 908.231587] env[62952]: _type = "Task" [ 908.231587] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.241093] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367134, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.295903] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.314875] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.316016] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.319500] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.589s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.320773] env[62952]: INFO nova.compute.claims [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.419460] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.419740] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.419953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.420159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.420325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.422752] env[62952]: INFO nova.compute.manager [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Terminating instance [ 908.427040] env[62952]: DEBUG nova.compute.manager [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.427040] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.427886] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b934b2fd-c528-4de8-843a-42afa751377b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.436053] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.436320] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-579eede6-d23f-4c8d-8844-c72221865007 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.443125] env[62952]: DEBUG oslo_vmware.api [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 908.443125] env[62952]: value = "task-1367135" [ 908.443125] env[62952]: _type = "Task" [ 908.443125] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.452341] env[62952]: DEBUG oslo_vmware.api [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.743026] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367134, 'name': PowerOnVM_Task} progress is 96%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.818011] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.818573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.818784] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.824401] env[62952]: DEBUG nova.compute.utils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.828119] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.828314] env[62952]: DEBUG nova.network.neutron [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.954089] env[62952]: DEBUG oslo_vmware.api [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367135, 'name': PowerOffVM_Task, 'duration_secs': 0.183415} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.954089] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 908.954089] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 908.954089] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-275b116f-9275-4d5a-9225-a750494f8faf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.016908] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.017143] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.017312] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Deleting the datastore file [datastore1] d2faf1e0-cfd1-4d87-ba77-0af92dc16643 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.017575] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccc51637-5813-46e9-a80d-8970f2fa998f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.022902] env[62952]: DEBUG nova.policy [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0194e5d37e4945ad7f2436032d2b60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6d1dbb357954f40995c72695f446232', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.025675] env[62952]: DEBUG oslo_vmware.api [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for the task: (returnval){ [ 909.025675] env[62952]: value = "task-1367137" [ 909.025675] env[62952]: _type = "Task" [ 909.025675] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.033899] env[62952]: DEBUG oslo_vmware.api [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.204381] env[62952]: INFO nova.scheduler.client.report [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Deleted allocations for instance 56aa8186-2bcb-47fc-a832-31654cbf218b [ 909.241832] env[62952]: DEBUG oslo_vmware.api [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367134, 'name': PowerOnVM_Task, 'duration_secs': 0.548063} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.243920] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.244319] env[62952]: INFO nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Took 7.96 seconds to spawn the instance on the hypervisor. [ 909.244546] env[62952]: DEBUG nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.245344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbabc50-3da0-4479-bf02-a3a0c25c458c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.327028] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.336543] env[62952]: DEBUG nova.network.neutron [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Successfully created port: 92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.536708] env[62952]: DEBUG oslo_vmware.api [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Task: {'id': task-1367137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199597} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.538864] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.539063] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 909.539463] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 909.539463] env[62952]: INFO nova.compute.manager [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Took 1.11 seconds to destroy the instance on the hypervisor. [ 909.539697] env[62952]: DEBUG oslo.service.loopingcall [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.540068] env[62952]: DEBUG nova.compute.manager [-] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 909.540189] env[62952]: DEBUG nova.network.neutron [-] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.570093] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25bafe22-2ecf-441f-b4db-5a117ba6bbcf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.578838] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf8f6e7-de74-47b7-a54e-ce7fe0354aa6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.612259] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76926bc2-c8a8-4bc2-b406-fb6bdf6b15f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.619382] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e902c23c-8f55-4217-b3ad-ca5f29aefe65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.632383] env[62952]: DEBUG nova.compute.provider_tree [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.713902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-659cc637-c9fe-4c5d-a9ea-5a52d58280b5 tempest-ServerRescueTestJSONUnderV235-1905484932 tempest-ServerRescueTestJSONUnderV235-1905484932-project-member] Lock "56aa8186-2bcb-47fc-a832-31654cbf218b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.553s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.763265] env[62952]: INFO nova.compute.manager [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Took 29.88 seconds to build instance. [ 909.863960] env[62952]: DEBUG nova.compute.manager [req-51fef963-e237-469b-bc44-2f3f35e2f5cf req-ac17dca6-6eb4-4e2f-b072-232abe1f1fbd service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Received event network-vif-deleted-e99d8558-48fe-47e0-9534-c7953ad6fa8c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.864174] env[62952]: INFO nova.compute.manager [req-51fef963-e237-469b-bc44-2f3f35e2f5cf req-ac17dca6-6eb4-4e2f-b072-232abe1f1fbd service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Neutron deleted interface e99d8558-48fe-47e0-9534-c7953ad6fa8c; detaching it from the instance and deleting it from the info cache [ 909.864350] env[62952]: DEBUG nova.network.neutron [req-51fef963-e237-469b-bc44-2f3f35e2f5cf req-ac17dca6-6eb4-4e2f-b072-232abe1f1fbd service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.137279] env[62952]: DEBUG nova.scheduler.client.report [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.220038] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.267672] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db0c2acf-1337-410a-be2f-63dc1b5edfd0 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.424s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.339939] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.342391] env[62952]: DEBUG nova.network.neutron [-] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.367633] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32869329-50ba-40bb-92dd-10e7a278062a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.374403] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.374640] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.374815] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.375033] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.375190] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.375340] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.375541] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.375696] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.375859] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.376030] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.376207] env[62952]: DEBUG nova.virt.hardware [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.377284] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11edc451-221a-49f5-8360-be26438009eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.386665] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074ef854-e738-4cfb-9571-9396c082c0f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.393422] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b181250-ba0f-48f4-887d-c905f841e238 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.418101] env[62952]: DEBUG nova.compute.manager [req-51fef963-e237-469b-bc44-2f3f35e2f5cf req-ac17dca6-6eb4-4e2f-b072-232abe1f1fbd service nova] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Detach interface failed, port_id=e99d8558-48fe-47e0-9534-c7953ad6fa8c, reason: Instance d2faf1e0-cfd1-4d87-ba77-0af92dc16643 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 910.645020] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.645020] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.646135] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.281s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.647752] env[62952]: INFO nova.compute.claims [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.745905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.769106] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.845331] env[62952]: INFO nova.compute.manager [-] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Took 1.31 seconds to deallocate network for instance. [ 910.945310] env[62952]: DEBUG nova.network.neutron [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Successfully updated port: 92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.152315] env[62952]: DEBUG nova.compute.utils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.159920] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.160122] env[62952]: DEBUG nova.network.neutron [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.198041] env[62952]: DEBUG nova.policy [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0194e5d37e4945ad7f2436032d2b60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6d1dbb357954f40995c72695f446232', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.301027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.353707] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.449149] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "refresh_cache-802fd770-38aa-4417-b25e-fd83510bceef" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.449320] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "refresh_cache-802fd770-38aa-4417-b25e-fd83510bceef" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.449473] env[62952]: DEBUG nova.network.neutron [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.579439] env[62952]: DEBUG nova.network.neutron [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Successfully created port: a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.663577] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.892561] env[62952]: DEBUG nova.compute.manager [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Received event network-vif-plugged-92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.892772] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] Acquiring lock "802fd770-38aa-4417-b25e-fd83510bceef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.892980] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] Lock "802fd770-38aa-4417-b25e-fd83510bceef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.893279] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] Lock "802fd770-38aa-4417-b25e-fd83510bceef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.893385] env[62952]: DEBUG nova.compute.manager [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] No waiting events found dispatching network-vif-plugged-92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.893730] env[62952]: WARNING nova.compute.manager [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Received unexpected event network-vif-plugged-92eac1b0-516e-40aa-8e5b-5dc9299025fc for instance with vm_state building and task_state spawning. [ 911.893907] env[62952]: DEBUG nova.compute.manager [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Received event network-changed-92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.894078] env[62952]: DEBUG nova.compute.manager [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Refreshing instance network info cache due to event network-changed-92eac1b0-516e-40aa-8e5b-5dc9299025fc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.895037] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] Acquiring lock "refresh_cache-802fd770-38aa-4417-b25e-fd83510bceef" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.972508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4161d79-2b52-4da3-9918-4a6191b68fe8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.981427] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7005972-1268-4d75-86da-d45c33394fbd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.988894] env[62952]: DEBUG nova.network.neutron [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.014192] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c4de86-1bf5-40c1-bab9-2953302844f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.021709] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82f83a6-c661-4796-9002-f1797d07c5ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.036200] env[62952]: DEBUG nova.compute.provider_tree [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.236067] env[62952]: DEBUG nova.network.neutron [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Updating instance_info_cache with network_info: [{"id": "92eac1b0-516e-40aa-8e5b-5dc9299025fc", "address": "fa:16:3e:28:0b:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92eac1b0-51", "ovs_interfaceid": "92eac1b0-516e-40aa-8e5b-5dc9299025fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.545787] env[62952]: DEBUG nova.scheduler.client.report [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.679128] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.706420] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.706731] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.706936] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.707173] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.707356] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.707543] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.707786] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.708028] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.708260] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.708463] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.708680] env[62952]: DEBUG nova.virt.hardware [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.709699] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400da123-3fd2-4f46-884c-83d767d85186 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.723872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25465b2a-9311-4376-aaaf-2121f078250d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.740017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "refresh_cache-802fd770-38aa-4417-b25e-fd83510bceef" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.740986] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Instance network_info: |[{"id": "92eac1b0-516e-40aa-8e5b-5dc9299025fc", "address": "fa:16:3e:28:0b:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92eac1b0-51", "ovs_interfaceid": "92eac1b0-516e-40aa-8e5b-5dc9299025fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.741553] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] Acquired lock "refresh_cache-802fd770-38aa-4417-b25e-fd83510bceef" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.741777] env[62952]: DEBUG nova.network.neutron [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Refreshing network info cache for port 92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.742985] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:0b:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c0d5204b-f60e-4830-84c8-2fe246c28202', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92eac1b0-516e-40aa-8e5b-5dc9299025fc', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.750675] env[62952]: DEBUG oslo.service.loopingcall [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.751258] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.752515] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6c7ec924-d357-483c-a8cf-44eb30f1ad4e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.778886] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.778886] env[62952]: value = "task-1367138" [ 912.778886] env[62952]: _type = "Task" [ 912.778886] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.787507] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367138, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.047769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.048397] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.052260] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.700s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.054331] env[62952]: INFO nova.compute.claims [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.217426] env[62952]: DEBUG nova.network.neutron [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Successfully updated port: a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.289398] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367138, 'name': CreateVM_Task, 'duration_secs': 0.335443} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.289590] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.290286] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.290724] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.291103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.291952] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39b65dbf-0dea-44f2-a5b3-dffaeb2432b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.296974] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 913.296974] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]521caf25-1b61-b2ae-9aaf-64485b8fa135" [ 913.296974] env[62952]: _type = "Task" [ 913.296974] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.307601] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]521caf25-1b61-b2ae-9aaf-64485b8fa135, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.559382] env[62952]: DEBUG nova.compute.utils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.562852] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.563032] env[62952]: DEBUG nova.network.neutron [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.593081] env[62952]: DEBUG nova.network.neutron [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Updated VIF entry in instance network info cache for port 92eac1b0-516e-40aa-8e5b-5dc9299025fc. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 913.593477] env[62952]: DEBUG nova.network.neutron [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Updating instance_info_cache with network_info: [{"id": "92eac1b0-516e-40aa-8e5b-5dc9299025fc", "address": "fa:16:3e:28:0b:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92eac1b0-51", "ovs_interfaceid": "92eac1b0-516e-40aa-8e5b-5dc9299025fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.648168] env[62952]: DEBUG nova.policy [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '913eacf6a94845b189e0bd5df408138a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8c2a296e8fe4805b9bb761f91a407f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.719857] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "refresh_cache-2f475040-28df-47e4-bfc4-4c843a57885f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.720154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "refresh_cache-2f475040-28df-47e4-bfc4-4c843a57885f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.720154] env[62952]: DEBUG nova.network.neutron [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.813991] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]521caf25-1b61-b2ae-9aaf-64485b8fa135, 'name': SearchDatastore_Task, 'duration_secs': 0.039179} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.814318] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.814620] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.814797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.814948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.815144] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.815907] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9ecb27e-a046-41e9-b0fc-721c8f376c27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.823414] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.823697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.824303] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b95d36e7-f11e-4c5f-8d35-14c64f90e356 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.829327] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 913.829327] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5265fdb8-f461-9d75-a324-6b12d736cb72" [ 913.829327] env[62952]: _type = "Task" [ 913.829327] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.837318] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5265fdb8-f461-9d75-a324-6b12d736cb72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.063934] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.099613] env[62952]: DEBUG oslo_concurrency.lockutils [req-a4a3a765-f735-44d7-a50e-74207e2b0e7f req-79160b73-c40a-4890-b906-ebeea732d1b4 service nova] Releasing lock "refresh_cache-802fd770-38aa-4417-b25e-fd83510bceef" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.106592] env[62952]: DEBUG nova.compute.manager [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Received event network-vif-plugged-a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.106808] env[62952]: DEBUG oslo_concurrency.lockutils [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] Acquiring lock "2f475040-28df-47e4-bfc4-4c843a57885f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.107019] env[62952]: DEBUG oslo_concurrency.lockutils [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] Lock "2f475040-28df-47e4-bfc4-4c843a57885f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.107188] env[62952]: DEBUG oslo_concurrency.lockutils [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] Lock "2f475040-28df-47e4-bfc4-4c843a57885f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.108371] env[62952]: DEBUG nova.compute.manager [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] No waiting events found dispatching network-vif-plugged-a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.108371] env[62952]: WARNING nova.compute.manager [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Received unexpected event network-vif-plugged-a8d09359-772d-44c6-b657-8de81288140f for instance with vm_state building and task_state spawning. [ 914.108371] env[62952]: DEBUG nova.compute.manager [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Received event network-changed-a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.108371] env[62952]: DEBUG nova.compute.manager [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Refreshing instance network info cache due to event network-changed-a8d09359-772d-44c6-b657-8de81288140f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.108371] env[62952]: DEBUG oslo_concurrency.lockutils [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] Acquiring lock "refresh_cache-2f475040-28df-47e4-bfc4-4c843a57885f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.288726] env[62952]: DEBUG nova.network.neutron [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.341931] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5265fdb8-f461-9d75-a324-6b12d736cb72, 'name': SearchDatastore_Task, 'duration_secs': 0.011707} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.341931] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-767ce666-451f-4eab-aa9d-078a201aba28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.347534] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 914.347534] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a4e38b-213e-9bdf-16ed-a95071c9d997" [ 914.347534] env[62952]: _type = "Task" [ 914.347534] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.356387] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a4e38b-213e-9bdf-16ed-a95071c9d997, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.384701] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdebf50-e9b7-411d-9331-69a5251f3c65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.391955] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70edb73-550c-4521-9730-5d08d5255bfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.422675] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd66472-a286-4bdb-8359-501d97ca55ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.431071] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e63b68-859a-4fce-9297-fa3568acec0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.446909] env[62952]: DEBUG nova.compute.provider_tree [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.451206] env[62952]: DEBUG nova.network.neutron [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Successfully created port: f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.672269] env[62952]: DEBUG nova.network.neutron [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Updating instance_info_cache with network_info: [{"id": "a8d09359-772d-44c6-b657-8de81288140f", "address": "fa:16:3e:5b:8e:b7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8d09359-77", "ovs_interfaceid": "a8d09359-772d-44c6-b657-8de81288140f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.859997] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a4e38b-213e-9bdf-16ed-a95071c9d997, 'name': SearchDatastore_Task, 'duration_secs': 0.009833} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.860329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.860795] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 802fd770-38aa-4417-b25e-fd83510bceef/802fd770-38aa-4417-b25e-fd83510bceef.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.861080] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a81b2dd6-6e11-4b47-9886-4204d62d8c30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.868243] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 914.868243] env[62952]: value = "task-1367139" [ 914.868243] env[62952]: _type = "Task" [ 914.868243] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.876705] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.954644] env[62952]: DEBUG nova.scheduler.client.report [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.076017] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.107503] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.107684] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.107752] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.107911] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.108151] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.108322] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.108538] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.108700] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.108867] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.109043] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.109390] env[62952]: DEBUG nova.virt.hardware [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.110567] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8504cc-1546-4ecd-af92-0493b943a3f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.120826] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285ec238-db75-4df0-8ca4-6433abee302b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.175253] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "refresh_cache-2f475040-28df-47e4-bfc4-4c843a57885f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.175601] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Instance network_info: |[{"id": "a8d09359-772d-44c6-b657-8de81288140f", "address": "fa:16:3e:5b:8e:b7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8d09359-77", "ovs_interfaceid": "a8d09359-772d-44c6-b657-8de81288140f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.175930] env[62952]: DEBUG oslo_concurrency.lockutils [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] Acquired lock "refresh_cache-2f475040-28df-47e4-bfc4-4c843a57885f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.176164] env[62952]: DEBUG nova.network.neutron [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Refreshing network info cache for port a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.177590] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:8e:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c0d5204b-f60e-4830-84c8-2fe246c28202', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8d09359-772d-44c6-b657-8de81288140f', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.185772] env[62952]: DEBUG oslo.service.loopingcall [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.189402] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.189755] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fef9be4-9fe0-4749-a657-2e2388e74cda {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.219013] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.219013] env[62952]: value = "task-1367140" [ 915.219013] env[62952]: _type = "Task" [ 915.219013] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.229733] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367140, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.380184] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367139, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464327} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.380580] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 802fd770-38aa-4417-b25e-fd83510bceef/802fd770-38aa-4417-b25e-fd83510bceef.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.380927] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.381312] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fca13bdf-20fb-43f2-84df-3e1920d97e96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.388229] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 915.388229] env[62952]: value = "task-1367141" [ 915.388229] env[62952]: _type = "Task" [ 915.388229] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.396318] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.442150] env[62952]: DEBUG nova.network.neutron [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Updated VIF entry in instance network info cache for port a8d09359-772d-44c6-b657-8de81288140f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.442573] env[62952]: DEBUG nova.network.neutron [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Updating instance_info_cache with network_info: [{"id": "a8d09359-772d-44c6-b657-8de81288140f", "address": "fa:16:3e:5b:8e:b7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8d09359-77", "ovs_interfaceid": "a8d09359-772d-44c6-b657-8de81288140f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.460671] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.461260] env[62952]: DEBUG nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.464083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.517s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.465680] env[62952]: INFO nova.compute.claims [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.730309] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367140, 'name': CreateVM_Task, 'duration_secs': 0.32669} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.730694] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.731321] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.731562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.732096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.732620] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2125b7e8-0d04-4b61-b7af-26759eb5e4ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.737825] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 915.737825] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526c51c1-c53c-13b5-a99c-e2fe06e427fc" [ 915.737825] env[62952]: _type = "Task" [ 915.737825] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.748079] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526c51c1-c53c-13b5-a99c-e2fe06e427fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.899192] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066287} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.899591] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.900479] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efdfde2f-4369-49b0-8686-454b6ff97a06 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.924593] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 802fd770-38aa-4417-b25e-fd83510bceef/802fd770-38aa-4417-b25e-fd83510bceef.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.925177] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8761967-6cfb-4e71-9e5a-c7c395492544 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.946056] env[62952]: DEBUG oslo_concurrency.lockutils [req-dcd73a5b-1cfd-4912-8cf4-f24e28ecb488 req-42ded02a-161c-4986-a784-4a6b47c2aab2 service nova] Releasing lock "refresh_cache-2f475040-28df-47e4-bfc4-4c843a57885f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.947040] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 915.947040] env[62952]: value = "task-1367142" [ 915.947040] env[62952]: _type = "Task" [ 915.947040] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.954900] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367142, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.972810] env[62952]: DEBUG nova.compute.utils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.975990] env[62952]: DEBUG nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 916.199568] env[62952]: DEBUG nova.compute.manager [req-54cb884f-48af-4ce3-8351-4f6192e7d523 req-2288b88a-cb37-41f8-9a62-07e0821ad6b9 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Received event network-vif-plugged-f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.199801] env[62952]: DEBUG oslo_concurrency.lockutils [req-54cb884f-48af-4ce3-8351-4f6192e7d523 req-2288b88a-cb37-41f8-9a62-07e0821ad6b9 service nova] Acquiring lock "aef28168-98a7-4f65-80e7-731633339abf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.204080] env[62952]: DEBUG oslo_concurrency.lockutils [req-54cb884f-48af-4ce3-8351-4f6192e7d523 req-2288b88a-cb37-41f8-9a62-07e0821ad6b9 service nova] Lock "aef28168-98a7-4f65-80e7-731633339abf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.204080] env[62952]: DEBUG oslo_concurrency.lockutils [req-54cb884f-48af-4ce3-8351-4f6192e7d523 req-2288b88a-cb37-41f8-9a62-07e0821ad6b9 service nova] Lock "aef28168-98a7-4f65-80e7-731633339abf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.004s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.204080] env[62952]: DEBUG nova.compute.manager [req-54cb884f-48af-4ce3-8351-4f6192e7d523 req-2288b88a-cb37-41f8-9a62-07e0821ad6b9 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] No waiting events found dispatching network-vif-plugged-f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 916.204228] env[62952]: WARNING nova.compute.manager [req-54cb884f-48af-4ce3-8351-4f6192e7d523 req-2288b88a-cb37-41f8-9a62-07e0821ad6b9 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Received unexpected event network-vif-plugged-f80edb6b-5326-4531-a02f-366484e828ce for instance with vm_state building and task_state spawning. [ 916.253196] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526c51c1-c53c-13b5-a99c-e2fe06e427fc, 'name': SearchDatastore_Task, 'duration_secs': 0.014597} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.253486] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.253719] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.253949] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.254135] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.254348] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.254593] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc18d51d-cd74-45af-ac3b-da3f674eaee0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.267030] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.267030] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.267030] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5299ba8b-9fe3-4847-a126-3fdbe79fa250 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.272035] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 916.272035] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c647c6-ebff-56e3-6e30-a80e066bddce" [ 916.272035] env[62952]: _type = "Task" [ 916.272035] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.279794] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c647c6-ebff-56e3-6e30-a80e066bddce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.294633] env[62952]: DEBUG nova.network.neutron [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Successfully updated port: f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.384873] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.385117] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.458924] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367142, 'name': ReconfigVM_Task, 'duration_secs': 0.275337} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.459207] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 802fd770-38aa-4417-b25e-fd83510bceef/802fd770-38aa-4417-b25e-fd83510bceef.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.459835] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e81ab7a2-95aa-4723-84df-c3b2ec8b3d57 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.466702] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 916.466702] env[62952]: value = "task-1367143" [ 916.466702] env[62952]: _type = "Task" [ 916.466702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.475136] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367143, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.476698] env[62952]: DEBUG nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.706337] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7516d94-e9bb-48f2-ace6-e470c9cdfaa8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.713605] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96271486-2f48-405a-a59f-5527bbbd6fc4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.742670] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668e53eb-2863-4079-81e0-2041dfed230d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.749689] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7e0aeb-d301-4f46-af73-12beadd7e85e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.762235] env[62952]: DEBUG nova.compute.provider_tree [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 916.781676] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c647c6-ebff-56e3-6e30-a80e066bddce, 'name': SearchDatastore_Task, 'duration_secs': 0.03525} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.782439] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9ac8d8b-1f30-4c70-aacc-b3a44dc687eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.787412] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 916.787412] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52d114ed-f6d9-4531-5369-5bd1b354a92d" [ 916.787412] env[62952]: _type = "Task" [ 916.787412] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.797255] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.797425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.797576] env[62952]: DEBUG nova.network.neutron [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.798592] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d114ed-f6d9-4531-5369-5bd1b354a92d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.978230] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367143, 'name': Rename_Task, 'duration_secs': 0.142908} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.978230] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 916.978230] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02264f82-61ae-4899-bf49-3deb73c0bea2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.986362] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 916.986362] env[62952]: value = "task-1367144" [ 916.986362] env[62952]: _type = "Task" [ 916.986362] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.995058] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367144, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.284541] env[62952]: ERROR nova.scheduler.client.report [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [req-ac0825b0-521b-4c91-9923-8fd7f4357a32] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9b9257d4-3be7-42a5-ae8c-67cf64823681. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ac0825b0-521b-4c91-9923-8fd7f4357a32"}]} [ 917.296842] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d114ed-f6d9-4531-5369-5bd1b354a92d, 'name': SearchDatastore_Task, 'duration_secs': 0.011797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.297113] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.297368] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 2f475040-28df-47e4-bfc4-4c843a57885f/2f475040-28df-47e4-bfc4-4c843a57885f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.297626] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0df48a10-2572-4acf-b614-76d8b49c629b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.300086] env[62952]: DEBUG nova.scheduler.client.report [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 917.308103] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 917.308103] env[62952]: value = "task-1367145" [ 917.308103] env[62952]: _type = "Task" [ 917.308103] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.315920] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.316812] env[62952]: DEBUG nova.scheduler.client.report [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 917.317021] env[62952]: DEBUG nova.compute.provider_tree [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.328675] env[62952]: DEBUG nova.scheduler.client.report [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 917.333053] env[62952]: DEBUG nova.network.neutron [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.350150] env[62952]: DEBUG nova.scheduler.client.report [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 917.488395] env[62952]: DEBUG nova.network.neutron [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.490369] env[62952]: DEBUG nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.505036] env[62952]: DEBUG oslo_vmware.api [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367144, 'name': PowerOnVM_Task, 'duration_secs': 0.486445} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.505036] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.505245] env[62952]: INFO nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Took 7.17 seconds to spawn the instance on the hypervisor. [ 917.505418] env[62952]: DEBUG nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.506260] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bae1c1b-9b75-4745-9451-87676d6f3b39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.525883] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.527083] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.527083] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.527083] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.527083] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.527083] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.527313] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.527313] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.527313] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.527447] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.528674] env[62952]: DEBUG nova.virt.hardware [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.528759] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869f5105-a195-4e78-b4c1-e6cc459dfcb5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.537880] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193fd43a-2486-41e5-bcaa-cc2948a159f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.556108] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.562057] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Creating folder: Project (62fdcd361a5f46bf8f63f9aab22895c0). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.564746] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ae24b6a-6c43-42ff-a4f2-6333566492ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.579917] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Created folder: Project (62fdcd361a5f46bf8f63f9aab22895c0) in parent group-v290852. [ 917.579917] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Creating folder: Instances. Parent ref: group-v290884. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.582136] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d9023cb-52c2-4209-884f-3be5c1d36b4c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.599307] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Created folder: Instances in parent group-v290884. [ 917.599307] env[62952]: DEBUG oslo.service.loopingcall [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.599444] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.599731] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df8d46c1-02aa-4f73-b328-7ee06592615f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.619781] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.619781] env[62952]: value = "task-1367148" [ 917.619781] env[62952]: _type = "Task" [ 917.619781] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.631061] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367148, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.656770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d81d9f-2c98-4f62-8ae7-c9257fdede2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.669123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e00d45a-01bf-4590-a658-54792ff1e5bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.733208] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c2b2ee-9614-4d31-a8ae-d3a7945d39e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.749826] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634da0b8-cd7b-4ac3-a878-0a3020b10256 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.767613] env[62952]: DEBUG nova.compute.provider_tree [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.821773] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367145, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.993728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.994221] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Instance network_info: |[{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.994675] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:09:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f80edb6b-5326-4531-a02f-366484e828ce', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.002518] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating folder: Project (b8c2a296e8fe4805b9bb761f91a407f9). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 918.002871] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4293ca2-ee12-4a8b-b00d-ccc4f9d6db14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.013900] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created folder: Project (b8c2a296e8fe4805b9bb761f91a407f9) in parent group-v290852. [ 918.014131] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating folder: Instances. Parent ref: group-v290887. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 918.014362] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0af8c496-4b14-410b-982f-281d1841a876 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.025371] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created folder: Instances in parent group-v290887. [ 918.025779] env[62952]: DEBUG oslo.service.loopingcall [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.026019] env[62952]: INFO nova.compute.manager [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Took 28.81 seconds to build instance. [ 918.026838] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aef28168-98a7-4f65-80e7-731633339abf] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.027051] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3af36a85-f222-40a9-8755-6a4d336cd971 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.048209] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.048209] env[62952]: value = "task-1367151" [ 918.048209] env[62952]: _type = "Task" [ 918.048209] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.056264] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367151, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.130052] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367148, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.276202] env[62952]: DEBUG nova.compute.manager [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Received event network-changed-f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.276365] env[62952]: DEBUG nova.compute.manager [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Refreshing instance network info cache due to event network-changed-f80edb6b-5326-4531-a02f-366484e828ce. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.276836] env[62952]: DEBUG oslo_concurrency.lockutils [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] Acquiring lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.276836] env[62952]: DEBUG oslo_concurrency.lockutils [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] Acquired lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.277069] env[62952]: DEBUG nova.network.neutron [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Refreshing network info cache for port f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.298035] env[62952]: DEBUG nova.scheduler.client.report [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updated inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 918.298035] env[62952]: DEBUG nova.compute.provider_tree [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updating resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 generation from 93 to 94 during operation: update_inventory {{(pid=62952) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 918.298238] env[62952]: DEBUG nova.compute.provider_tree [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.321406] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.321689] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 2f475040-28df-47e4-bfc4-4c843a57885f/2f475040-28df-47e4-bfc4-4c843a57885f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.321919] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.322188] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f53b784-27f2-486a-96c6-99349d578234 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.328335] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 918.328335] env[62952]: value = "task-1367152" [ 918.328335] env[62952]: _type = "Task" [ 918.328335] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.335827] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367152, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.543599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-08bac2c1-1cd8-48d9-b289-c5cc3bd1b9ff tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "802fd770-38aa-4417-b25e-fd83510bceef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.484s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.558765] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367151, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.630277] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367148, 'name': CreateVM_Task, 'duration_secs': 0.722094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.630444] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.630885] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.631062] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.631380] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.631619] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14778594-2b3c-4755-a296-8c3001631b14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.635700] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 918.635700] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]522e47cc-6ed1-8f75-cc5a-7e683813c18c" [ 918.635700] env[62952]: _type = "Task" [ 918.635700] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.642773] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522e47cc-6ed1-8f75-cc5a-7e683813c18c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.803098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.339s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.803693] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.806386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.906s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.807793] env[62952]: INFO nova.compute.claims [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.839032] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367152, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.139243} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.839734] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.840216] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a8a730-e768-4d25-8747-565a8cc1e9ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.863788] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 2f475040-28df-47e4-bfc4-4c843a57885f/2f475040-28df-47e4-bfc4-4c843a57885f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.866567] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67f9be1a-47b1-4fae-8503-ea0f2b29b76d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.886411] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 918.886411] env[62952]: value = "task-1367153" [ 918.886411] env[62952]: _type = "Task" [ 918.886411] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.894836] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367153, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.028638] env[62952]: DEBUG nova.network.neutron [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updated VIF entry in instance network info cache for port f80edb6b-5326-4531-a02f-366484e828ce. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.029062] env[62952]: DEBUG nova.network.neutron [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.045814] env[62952]: DEBUG nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.058437] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367151, 'name': CreateVM_Task, 'duration_secs': 0.874101} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.058685] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aef28168-98a7-4f65-80e7-731633339abf] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.059394] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.145485] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522e47cc-6ed1-8f75-cc5a-7e683813c18c, 'name': SearchDatastore_Task, 'duration_secs': 0.008744} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.145779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.146484] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.146484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.146484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.146650] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.146861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.147180] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.147413] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7138486-2055-4a91-bdb1-27374c6717d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.149105] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d067e8da-b539-497f-bd17-d81babe618df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.155184] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 919.155184] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]521c9f0a-f0a7-9372-057a-26f3c2357f20" [ 919.155184] env[62952]: _type = "Task" [ 919.155184] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.159196] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.159373] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.160585] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c56ba1d4-78ab-43c7-b59a-86fb60f96120 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.165484] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]521c9f0a-f0a7-9372-057a-26f3c2357f20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.168118] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 919.168118] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]521e578a-6a67-6f35-8fe7-9f2e2590a90a" [ 919.168118] env[62952]: _type = "Task" [ 919.168118] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.175200] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]521e578a-6a67-6f35-8fe7-9f2e2590a90a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.312604] env[62952]: DEBUG nova.compute.utils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 919.313631] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 919.313925] env[62952]: DEBUG nova.network.neutron [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 919.367968] env[62952]: DEBUG nova.policy [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17b50c91ccd347e4b45126a8f23c0278', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9657a113032417fb4b97716a772f8ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 919.395937] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367153, 'name': ReconfigVM_Task, 'duration_secs': 0.248356} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.396214] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 2f475040-28df-47e4-bfc4-4c843a57885f/2f475040-28df-47e4-bfc4-4c843a57885f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.396804] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6373bb8c-1947-488e-a7c7-c4cc7a988ce2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.403094] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 919.403094] env[62952]: value = "task-1367154" [ 919.403094] env[62952]: _type = "Task" [ 919.403094] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.410872] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367154, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.531453] env[62952]: DEBUG oslo_concurrency.lockutils [req-83c956c0-58ec-4ecf-9de3-6d235c964660 req-341c9a6e-4930-42bf-9985-964e738b812a service nova] Releasing lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.568692] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.605757] env[62952]: DEBUG nova.network.neutron [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Successfully created port: feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.665668] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]521c9f0a-f0a7-9372-057a-26f3c2357f20, 'name': SearchDatastore_Task, 'duration_secs': 0.00867} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.665950] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.666189] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.666401] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.676430] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]521e578a-6a67-6f35-8fe7-9f2e2590a90a, 'name': SearchDatastore_Task, 'duration_secs': 0.008987} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.677011] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a66eae7-d951-4a2b-b4e1-ed2cd27e3d5c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.681458] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 919.681458] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5268cd2d-e876-727f-d4a3-f549ec6c60f0" [ 919.681458] env[62952]: _type = "Task" [ 919.681458] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.688642] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5268cd2d-e876-727f-d4a3-f549ec6c60f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.818895] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.916333] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367154, 'name': Rename_Task, 'duration_secs': 0.136433} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.916771] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.917165] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3989f013-b7ba-4934-9995-6bae9b7ddc5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.924815] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 919.924815] env[62952]: value = "task-1367155" [ 919.924815] env[62952]: _type = "Task" [ 919.924815] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.932774] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.061609] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25364607-1218-46b2-927a-c96e6f1925fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.069754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b42517-d54e-4a9e-9794-f8cfa9953f9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.103697] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3dd099-4c8e-4750-bdff-dc4139a9f268 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.112023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21dc3d48-563e-4a6b-8578-e7aba2227b03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.127097] env[62952]: DEBUG nova.compute.provider_tree [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.192095] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5268cd2d-e876-727f-d4a3-f549ec6c60f0, 'name': SearchDatastore_Task, 'duration_secs': 0.009176} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.192419] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.192709] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 5914a35a-0934-4f8f-81e8-d91bc690a9cb/5914a35a-0934-4f8f-81e8-d91bc690a9cb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.193714] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.193714] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.193714] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82270a05-7aa8-4812-9e63-4c772eb587bc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.195596] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3adfcf3f-ad06-4a8a-aa9c-b75a6c239e78 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.203866] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.204078] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.205896] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-654c1a7d-e806-4e96-8a93-bc52524d7674 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.208788] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 920.208788] env[62952]: value = "task-1367156" [ 920.208788] env[62952]: _type = "Task" [ 920.208788] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.213755] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 920.213755] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524eac27-f0d8-8a34-03ef-c53974afc55a" [ 920.213755] env[62952]: _type = "Task" [ 920.213755] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.220819] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.226177] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524eac27-f0d8-8a34-03ef-c53974afc55a, 'name': SearchDatastore_Task, 'duration_secs': 0.009094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.226976] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7251ebdb-9fb2-40ae-8025-8b1db1bb8275 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.231967] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 920.231967] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5233b929-699a-377a-fe90-35c57f62cdf9" [ 920.231967] env[62952]: _type = "Task" [ 920.231967] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.239913] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5233b929-699a-377a-fe90-35c57f62cdf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.438273] env[62952]: DEBUG oslo_vmware.api [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367155, 'name': PowerOnVM_Task, 'duration_secs': 0.499571} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.438273] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.438273] env[62952]: INFO nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Took 7.76 seconds to spawn the instance on the hypervisor. [ 920.438273] env[62952]: DEBUG nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.439343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265b021b-3df3-4ff9-9d17-d15a57a4b0f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.630452] env[62952]: DEBUG nova.scheduler.client.report [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.719070] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367156, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.742542] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5233b929-699a-377a-fe90-35c57f62cdf9, 'name': SearchDatastore_Task, 'duration_secs': 0.008822} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.742818] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.743085] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] aef28168-98a7-4f65-80e7-731633339abf/aef28168-98a7-4f65-80e7-731633339abf.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.743343] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce7ea190-ac42-45b3-bd00-0c0b3429f26a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.749394] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 920.749394] env[62952]: value = "task-1367157" [ 920.749394] env[62952]: _type = "Task" [ 920.749394] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.760235] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367157, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.831070] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.855584] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.855831] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.855989] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.856184] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.856330] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.856474] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.856680] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.856839] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.857008] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.857178] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.857352] env[62952]: DEBUG nova.virt.hardware [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.858215] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80a92db-3b52-46fc-8e19-33126b6d7c22 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.865779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee15db44-ebc5-4304-8f40-b27eaeb25b32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.958259] env[62952]: INFO nova.compute.manager [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Took 27.26 seconds to build instance. [ 921.135515] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.136095] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.141108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.150s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.141435] env[62952]: DEBUG nova.objects.instance [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lazy-loading 'resources' on Instance uuid 351b1672-ddb3-4562-9965-d65a7d869e82 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.205775] env[62952]: DEBUG nova.compute.manager [req-76607ab9-dd2f-46fa-8112-848d54f54eb6 req-45755f7b-44c8-4315-9c9e-243f8eaf732f service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Received event network-vif-plugged-feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.206039] env[62952]: DEBUG oslo_concurrency.lockutils [req-76607ab9-dd2f-46fa-8112-848d54f54eb6 req-45755f7b-44c8-4315-9c9e-243f8eaf732f service nova] Acquiring lock "34f5995f-692c-4f0d-8b15-0d388df7e34e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.206261] env[62952]: DEBUG oslo_concurrency.lockutils [req-76607ab9-dd2f-46fa-8112-848d54f54eb6 req-45755f7b-44c8-4315-9c9e-243f8eaf732f service nova] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.206435] env[62952]: DEBUG oslo_concurrency.lockutils [req-76607ab9-dd2f-46fa-8112-848d54f54eb6 req-45755f7b-44c8-4315-9c9e-243f8eaf732f service nova] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.206621] env[62952]: DEBUG nova.compute.manager [req-76607ab9-dd2f-46fa-8112-848d54f54eb6 req-45755f7b-44c8-4315-9c9e-243f8eaf732f service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] No waiting events found dispatching network-vif-plugged-feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.206789] env[62952]: WARNING nova.compute.manager [req-76607ab9-dd2f-46fa-8112-848d54f54eb6 req-45755f7b-44c8-4315-9c9e-243f8eaf732f service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Received unexpected event network-vif-plugged-feeba8e0-a5e3-4e76-9991-96695e8032eb for instance with vm_state building and task_state spawning. [ 921.223346] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532693} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.223462] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 5914a35a-0934-4f8f-81e8-d91bc690a9cb/5914a35a-0934-4f8f-81e8-d91bc690a9cb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.223678] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.224041] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db2cc70c-6360-414f-a3b2-89db15c384ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.231670] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 921.231670] env[62952]: value = "task-1367158" [ 921.231670] env[62952]: _type = "Task" [ 921.231670] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.239843] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367158, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.258705] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367157, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462588} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.258988] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] aef28168-98a7-4f65-80e7-731633339abf/aef28168-98a7-4f65-80e7-731633339abf.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.259240] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.259496] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8d47a53-46b3-4a15-8c48-0608c72e9848 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.266234] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 921.266234] env[62952]: value = "task-1367159" [ 921.266234] env[62952]: _type = "Task" [ 921.266234] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.273982] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367159, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.460985] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ec2a8b68-3326-49b4-ac4b-ff249a7565c4 tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "2f475040-28df-47e4-bfc4-4c843a57885f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.008s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.645295] env[62952]: DEBUG nova.compute.utils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.649588] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.649588] env[62952]: DEBUG nova.network.neutron [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 921.666448] env[62952]: DEBUG nova.network.neutron [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Successfully updated port: feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.713368] env[62952]: DEBUG nova.policy [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2579d36f477f4a94b3d839ed67108e5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '373ca19a1eb64177b43d5e107d192959', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 921.721181] env[62952]: DEBUG nova.compute.manager [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Received event network-changed-feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.721181] env[62952]: DEBUG nova.compute.manager [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Refreshing instance network info cache due to event network-changed-feeba8e0-a5e3-4e76-9991-96695e8032eb. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.721417] env[62952]: DEBUG oslo_concurrency.lockutils [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] Acquiring lock "refresh_cache-34f5995f-692c-4f0d-8b15-0d388df7e34e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.721590] env[62952]: DEBUG oslo_concurrency.lockutils [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] Acquired lock "refresh_cache-34f5995f-692c-4f0d-8b15-0d388df7e34e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.721773] env[62952]: DEBUG nova.network.neutron [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Refreshing network info cache for port feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.742087] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367158, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075205} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.742411] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.743323] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89d60cb-8ae2-499c-bf51-ea16bcd91e43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.769292] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 5914a35a-0934-4f8f-81e8-d91bc690a9cb/5914a35a-0934-4f8f-81e8-d91bc690a9cb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.773125] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8918a608-2193-4c13-adbd-dec8306786a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.795682] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367159, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.797025] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.797375] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 921.797375] env[62952]: value = "task-1367160" [ 921.797375] env[62952]: _type = "Task" [ 921.797375] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.800344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf49509d-50a7-41fb-b961-d9508e716026 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.825256] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] aef28168-98a7-4f65-80e7-731633339abf/aef28168-98a7-4f65-80e7-731633339abf.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.831016] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e8e5c51-a70b-4e7a-a4cd-d749350b9b24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.845720] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367160, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.851144] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 921.851144] env[62952]: value = "task-1367161" [ 921.851144] env[62952]: _type = "Task" [ 921.851144] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.861809] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.964846] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.973420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eae204-2779-4c16-a2f5-f8fdac021682 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.981092] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c7d8e5-5e23-445b-8f84-e2b4f3e257d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.017453] env[62952]: DEBUG nova.network.neutron [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Successfully created port: 1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.020555] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37f539d-73c5-4d0b-a059-6d40f042a2c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.029785] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cea8f8d-79e2-4b76-849d-4410e806c7a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.045344] env[62952]: DEBUG nova.compute.provider_tree [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.153575] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.171099] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-34f5995f-692c-4f0d-8b15-0d388df7e34e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.280251] env[62952]: DEBUG nova.network.neutron [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 922.313759] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367160, 'name': ReconfigVM_Task, 'duration_secs': 0.297029} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.314552] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 5914a35a-0934-4f8f-81e8-d91bc690a9cb/5914a35a-0934-4f8f-81e8-d91bc690a9cb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.317330] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12beff15-2df6-4801-83c1-fec943e72a67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.323292] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 922.323292] env[62952]: value = "task-1367162" [ 922.323292] env[62952]: _type = "Task" [ 922.323292] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.332986] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367162, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.362506] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367161, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.485404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.551155] env[62952]: DEBUG nova.scheduler.client.report [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.617837] env[62952]: DEBUG nova.network.neutron [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.836045] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367162, 'name': Rename_Task, 'duration_secs': 0.363932} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.836648] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.836949] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aad3e3dd-f776-44f7-bafa-72f199c596c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.846395] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 922.846395] env[62952]: value = "task-1367163" [ 922.846395] env[62952]: _type = "Task" [ 922.846395] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.854904] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367163, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.865793] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367161, 'name': ReconfigVM_Task, 'duration_secs': 0.853599} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.866226] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Reconfigured VM instance instance-0000003f to attach disk [datastore2] aef28168-98a7-4f65-80e7-731633339abf/aef28168-98a7-4f65-80e7-731633339abf.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.867114] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b19bfcd-f085-4136-887e-9e7a7d53dcaf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.875682] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 922.875682] env[62952]: value = "task-1367164" [ 922.875682] env[62952]: _type = "Task" [ 922.875682] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.885592] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367164, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.056440] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.915s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.059613] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.175s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.061249] env[62952]: INFO nova.compute.claims [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.079177] env[62952]: INFO nova.scheduler.client.report [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Deleted allocations for instance 351b1672-ddb3-4562-9965-d65a7d869e82 [ 923.120605] env[62952]: DEBUG oslo_concurrency.lockutils [req-2db9e5e0-15b5-4c52-ab86-67e52b6e5d5c req-9eccb8db-c962-4537-b20e-91ea3dccd7d8 service nova] Releasing lock "refresh_cache-34f5995f-692c-4f0d-8b15-0d388df7e34e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.121133] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-34f5995f-692c-4f0d-8b15-0d388df7e34e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.121325] env[62952]: DEBUG nova.network.neutron [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.164554] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.191085] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.191399] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.191501] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.191681] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.191827] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.191974] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.192214] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.192439] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.192534] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.192696] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.192868] env[62952]: DEBUG nova.virt.hardware [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.193818] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9f2ed5-6876-4a37-9f83-4a53f0438d84 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.202624] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5576b5-28e6-4224-8014-cc51e36ede7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.356424] env[62952]: DEBUG oslo_vmware.api [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367163, 'name': PowerOnVM_Task, 'duration_secs': 0.433445} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.356685] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.356882] env[62952]: INFO nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Took 5.87 seconds to spawn the instance on the hypervisor. [ 923.357507] env[62952]: DEBUG nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.358312] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0281a4e-7bf0-489b-b83c-c43d8778859a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.386792] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367164, 'name': Rename_Task, 'duration_secs': 0.158435} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.387099] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.387346] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4aff6fcb-17b6-4771-8592-0f7eacc7df56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.395086] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 923.395086] env[62952]: value = "task-1367165" [ 923.395086] env[62952]: _type = "Task" [ 923.395086] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.404530] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367165, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.586851] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a0b8fc40-5edd-4630-ab48-d316afd2d013 tempest-InstanceActionsTestJSON-726318605 tempest-InstanceActionsTestJSON-726318605-project-member] Lock "351b1672-ddb3-4562-9965-d65a7d869e82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.598s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.681802] env[62952]: DEBUG nova.network.neutron [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.878025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "a63ec2ac-4484-4360-962b-105a5338c1b8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.878025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.878521] env[62952]: DEBUG nova.compute.manager [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.881640] env[62952]: INFO nova.compute.manager [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Took 26.55 seconds to build instance. [ 923.887482] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6838a74b-b2cb-4cd3-8b89-2a2c0ec2d75d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.896762] env[62952]: DEBUG nova.compute.manager [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 923.897450] env[62952]: DEBUG nova.objects.instance [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lazy-loading 'flavor' on Instance uuid a63ec2ac-4484-4360-962b-105a5338c1b8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.911148] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367165, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.107223] env[62952]: DEBUG nova.network.neutron [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Updating instance_info_cache with network_info: [{"id": "feeba8e0-a5e3-4e76-9991-96695e8032eb", "address": "fa:16:3e:75:38:f8", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeeba8e0-a5", "ovs_interfaceid": "feeba8e0-a5e3-4e76-9991-96695e8032eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.337174] env[62952]: DEBUG nova.compute.manager [req-02ce57b8-8d3e-475a-a8c9-f7c2e0352321 req-9821bcec-5f8f-430e-a695-dc4ce5e0ee14 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Received event network-vif-plugged-1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.337398] env[62952]: DEBUG oslo_concurrency.lockutils [req-02ce57b8-8d3e-475a-a8c9-f7c2e0352321 req-9821bcec-5f8f-430e-a695-dc4ce5e0ee14 service nova] Acquiring lock "4983f354-77dc-4e84-969b-4c74441fa568-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.337606] env[62952]: DEBUG oslo_concurrency.lockutils [req-02ce57b8-8d3e-475a-a8c9-f7c2e0352321 req-9821bcec-5f8f-430e-a695-dc4ce5e0ee14 service nova] Lock "4983f354-77dc-4e84-969b-4c74441fa568-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.337773] env[62952]: DEBUG oslo_concurrency.lockutils [req-02ce57b8-8d3e-475a-a8c9-f7c2e0352321 req-9821bcec-5f8f-430e-a695-dc4ce5e0ee14 service nova] Lock "4983f354-77dc-4e84-969b-4c74441fa568-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.337968] env[62952]: DEBUG nova.compute.manager [req-02ce57b8-8d3e-475a-a8c9-f7c2e0352321 req-9821bcec-5f8f-430e-a695-dc4ce5e0ee14 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] No waiting events found dispatching network-vif-plugged-1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.338163] env[62952]: WARNING nova.compute.manager [req-02ce57b8-8d3e-475a-a8c9-f7c2e0352321 req-9821bcec-5f8f-430e-a695-dc4ce5e0ee14 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Received unexpected event network-vif-plugged-1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 for instance with vm_state building and task_state spawning. [ 924.390751] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2bd1db-4e89-4849-8a07-8086e56f3c8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.394316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-920dd7fc-22c7-4310-8ecc-82989e9a95c1 tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.550s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.402946] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f7317e-4daa-4fc8-b7f6-1a2204c4908b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.411696] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367165, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.412178] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.412932] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ab22e85-263e-45bf-b1d0-92224a0e4a63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.441104] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e61a4b4-626b-4a2d-9e2e-4c0042a3e809 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.444016] env[62952]: DEBUG oslo_vmware.api [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 924.444016] env[62952]: value = "task-1367166" [ 924.444016] env[62952]: _type = "Task" [ 924.444016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.463905] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081de1f0-dd26-47d3-ab5f-ad2a449c8800 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.468311] env[62952]: DEBUG oslo_vmware.api [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367166, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.480410] env[62952]: DEBUG nova.compute.provider_tree [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.610218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-34f5995f-692c-4f0d-8b15-0d388df7e34e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.610541] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance network_info: |[{"id": "feeba8e0-a5e3-4e76-9991-96695e8032eb", "address": "fa:16:3e:75:38:f8", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeeba8e0-a5", "ovs_interfaceid": "feeba8e0-a5e3-4e76-9991-96695e8032eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.610950] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:38:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'feeba8e0-a5e3-4e76-9991-96695e8032eb', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.618718] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating folder: Project (c9657a113032417fb4b97716a772f8ce). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.619008] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ea4c30c-44e4-4f67-9e79-c76e6a617f51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.630801] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created folder: Project (c9657a113032417fb4b97716a772f8ce) in parent group-v290852. [ 924.631065] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating folder: Instances. Parent ref: group-v290890. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.631249] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73df9e50-aa03-4393-b5f1-219d8e3cb99c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.640628] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created folder: Instances in parent group-v290890. [ 924.640865] env[62952]: DEBUG oslo.service.loopingcall [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.641071] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 924.641278] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e799436-df1e-4ebd-beab-fabf9b1d11db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.661761] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.661761] env[62952]: value = "task-1367169" [ 924.661761] env[62952]: _type = "Task" [ 924.661761] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.669131] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367169, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.732046] env[62952]: DEBUG nova.network.neutron [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Successfully updated port: 1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.760780] env[62952]: DEBUG nova.compute.manager [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Received event network-changed-1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.761152] env[62952]: DEBUG nova.compute.manager [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Refreshing instance network info cache due to event network-changed-1e875b4e-53fe-4d5e-b905-f34fcdcc8e86. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.761276] env[62952]: DEBUG oslo_concurrency.lockutils [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] Acquiring lock "refresh_cache-4983f354-77dc-4e84-969b-4c74441fa568" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.761363] env[62952]: DEBUG oslo_concurrency.lockutils [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] Acquired lock "refresh_cache-4983f354-77dc-4e84-969b-4c74441fa568" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.761497] env[62952]: DEBUG nova.network.neutron [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Refreshing network info cache for port 1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.823497] env[62952]: DEBUG nova.compute.manager [None req-b07d131f-7952-464a-b2f6-7f0cfe8cb81e tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.824491] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4b59cd-3cd1-47ce-916e-4a3d50d2251c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.898824] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.908981] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367165, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.954112] env[62952]: DEBUG oslo_vmware.api [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367166, 'name': PowerOffVM_Task, 'duration_secs': 0.359913} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.954539] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.954710] env[62952]: DEBUG nova.compute.manager [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.955506] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9f292b-6559-4205-a485-4b5bd938945c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.968762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.969300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.969620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.969878] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.970107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.972229] env[62952]: INFO nova.compute.manager [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Terminating instance [ 924.974077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "refresh_cache-5914a35a-0934-4f8f-81e8-d91bc690a9cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.974550] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquired lock "refresh_cache-5914a35a-0934-4f8f-81e8-d91bc690a9cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.974786] env[62952]: DEBUG nova.network.neutron [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.983198] env[62952]: DEBUG nova.scheduler.client.report [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.174995] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367169, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.235860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "refresh_cache-4983f354-77dc-4e84-969b-4c74441fa568" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.316795] env[62952]: DEBUG nova.network.neutron [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.337491] env[62952]: INFO nova.compute.manager [None req-b07d131f-7952-464a-b2f6-7f0cfe8cb81e tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] instance snapshotting [ 925.337491] env[62952]: DEBUG nova.objects.instance [None req-b07d131f-7952-464a-b2f6-7f0cfe8cb81e tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lazy-loading 'flavor' on Instance uuid 5914a35a-0934-4f8f-81e8-d91bc690a9cb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.414867] env[62952]: DEBUG oslo_vmware.api [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367165, 'name': PowerOnVM_Task, 'duration_secs': 1.59179} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.415143] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.415471] env[62952]: INFO nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Took 10.34 seconds to spawn the instance on the hypervisor. [ 925.415652] env[62952]: DEBUG nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.416734] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4777c0-4cbf-4782-a68b-4c8b662b8b2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.434538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.467516] env[62952]: DEBUG oslo_concurrency.lockutils [None req-de56bdc4-cb86-4186-b7f9-8132d81b794c tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.589s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.479430] env[62952]: DEBUG nova.network.neutron [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.488328] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.488674] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.491136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.673s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.492747] env[62952]: INFO nova.compute.claims [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.497643] env[62952]: DEBUG nova.network.neutron [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.554910] env[62952]: DEBUG nova.network.neutron [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.671598] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367169, 'name': CreateVM_Task, 'duration_secs': 0.548069} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.671775] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.672470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.672640] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.672961] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.673263] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c8523f5-cbc8-469f-a523-133e45ee1606 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.677684] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 925.677684] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5281aa0e-a366-6859-05e1-70e17195f0de" [ 925.677684] env[62952]: _type = "Task" [ 925.677684] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.686142] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5281aa0e-a366-6859-05e1-70e17195f0de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.843113] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5c155d-a6a6-4c3b-bd88-020d0e6c18bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.863191] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed35126-d83c-4ba9-bfc8-33d94fecb552 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.939380] env[62952]: INFO nova.compute.manager [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Took 30.60 seconds to build instance. [ 925.982634] env[62952]: DEBUG oslo_concurrency.lockutils [req-af38d2e1-e609-4b15-b2c2-8071343e773e req-a266620f-533f-4740-bcf1-f83f8407b5c8 service nova] Releasing lock "refresh_cache-4983f354-77dc-4e84-969b-4c74441fa568" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.983042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquired lock "refresh_cache-4983f354-77dc-4e84-969b-4c74441fa568" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.983226] env[62952]: DEBUG nova.network.neutron [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.998952] env[62952]: DEBUG nova.compute.utils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.001254] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.001448] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 926.060485] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Releasing lock "refresh_cache-5914a35a-0934-4f8f-81e8-d91bc690a9cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.060859] env[62952]: DEBUG nova.compute.manager [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.061094] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.062106] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7046631-5648-46e1-aeb8-b5615e2abf47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.066883] env[62952]: DEBUG nova.policy [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '378e043e429e4b3098e0003311d0b6de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f7b88e9cabf41a7802ebd6b3b84d51e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.072879] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.073174] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-209dcecc-f902-434e-84e9-00249d33f3ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.079748] env[62952]: DEBUG oslo_vmware.api [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 926.079748] env[62952]: value = "task-1367170" [ 926.079748] env[62952]: _type = "Task" [ 926.079748] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.090262] env[62952]: DEBUG oslo_vmware.api [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.190167] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5281aa0e-a366-6859-05e1-70e17195f0de, 'name': SearchDatastore_Task, 'duration_secs': 0.014144} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.190498] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.190751] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.191066] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.191239] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.191432] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.191712] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca769d79-67fb-4a8f-ae53-698f1b8633ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.200931] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.201187] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.201918] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-261f7492-7a1e-4c99-9357-56e51aa61b83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.210798] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 926.210798] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52828449-f953-0fbb-7766-558c4c6ef4be" [ 926.210798] env[62952]: _type = "Task" [ 926.210798] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.218650] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52828449-f953-0fbb-7766-558c4c6ef4be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.373608] env[62952]: DEBUG nova.compute.manager [None req-b07d131f-7952-464a-b2f6-7f0cfe8cb81e tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Instance disappeared during snapshot {{(pid=62952) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 926.429516] env[62952]: DEBUG nova.objects.instance [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lazy-loading 'flavor' on Instance uuid a63ec2ac-4484-4360-962b-105a5338c1b8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.441754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3f36794-e7df-4095-a7ba-8876b99ae35e tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "aef28168-98a7-4f65-80e7-731633339abf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.106s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.509754] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.545607] env[62952]: DEBUG nova.compute.manager [None req-b07d131f-7952-464a-b2f6-7f0cfe8cb81e tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Found 0 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 926.565282] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Successfully created port: c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.572250] env[62952]: DEBUG nova.network.neutron [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.590297] env[62952]: DEBUG oslo_vmware.api [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367170, 'name': PowerOffVM_Task, 'duration_secs': 0.150268} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.590884] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.591197] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.591505] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a914491-5a86-4b31-84b7-4eda15fce2da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.624321] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.624550] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.624735] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Deleting the datastore file [datastore2] 5914a35a-0934-4f8f-81e8-d91bc690a9cb {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.625046] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54c07017-ac02-4629-9d48-7b47959e27f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.632695] env[62952]: DEBUG oslo_vmware.api [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for the task: (returnval){ [ 926.632695] env[62952]: value = "task-1367172" [ 926.632695] env[62952]: _type = "Task" [ 926.632695] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.641442] env[62952]: DEBUG oslo_vmware.api [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.724375] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52828449-f953-0fbb-7766-558c4c6ef4be, 'name': SearchDatastore_Task, 'duration_secs': 0.025397} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.729171] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac083ac7-7e7e-458c-8048-0a71ca07bfcb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.735883] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 926.735883] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]529a34ce-7c87-58df-e947-7315ea55d049" [ 926.735883] env[62952]: _type = "Task" [ 926.735883] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.746599] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529a34ce-7c87-58df-e947-7315ea55d049, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.840332] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22768da4-51ac-4183-ad16-4eeabaa4bc8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.848502] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b4ceca-66bd-47f5-aa30-1b4be7a08628 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.885041] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553e3c25-b122-4683-af8d-8846f55fee0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.893509] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92319cb8-0c67-4878-acc4-1917ea34b585 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.907781] env[62952]: DEBUG nova.compute.provider_tree [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.935023] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.935208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquired lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.935382] env[62952]: DEBUG nova.network.neutron [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 926.935627] env[62952]: DEBUG nova.objects.instance [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lazy-loading 'info_cache' on Instance uuid a63ec2ac-4484-4360-962b-105a5338c1b8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.944682] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 926.980576] env[62952]: DEBUG nova.network.neutron [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Updating instance_info_cache with network_info: [{"id": "1e875b4e-53fe-4d5e-b905-f34fcdcc8e86", "address": "fa:16:3e:89:02:e4", "network": {"id": "b5311ec0-3267-40d6-90ef-119177df8283", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1457467599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "373ca19a1eb64177b43d5e107d192959", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e875b4e-53", "ovs_interfaceid": "1e875b4e-53fe-4d5e-b905-f34fcdcc8e86", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.145596] env[62952]: DEBUG oslo_vmware.api [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Task: {'id': task-1367172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201374} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.145844] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.146042] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.146229] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.146408] env[62952]: INFO nova.compute.manager [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Took 1.09 seconds to destroy the instance on the hypervisor. [ 927.146647] env[62952]: DEBUG oslo.service.loopingcall [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.146838] env[62952]: DEBUG nova.compute.manager [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.146931] env[62952]: DEBUG nova.network.neutron [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.256454] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529a34ce-7c87-58df-e947-7315ea55d049, 'name': SearchDatastore_Task, 'duration_secs': 0.012171} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.256780] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.257055] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 927.261057] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fbb2dbf-4873-4e04-9107-bc68dc97ab1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.268681] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 927.268681] env[62952]: value = "task-1367173" [ 927.268681] env[62952]: _type = "Task" [ 927.268681] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.281487] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367173, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.322461] env[62952]: DEBUG nova.network.neutron [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 927.411151] env[62952]: DEBUG nova.scheduler.client.report [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.417794] env[62952]: DEBUG nova.compute.manager [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Received event network-changed-f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.418107] env[62952]: DEBUG nova.compute.manager [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Refreshing instance network info cache due to event network-changed-f80edb6b-5326-4531-a02f-366484e828ce. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.418419] env[62952]: DEBUG oslo_concurrency.lockutils [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] Acquiring lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.418653] env[62952]: DEBUG oslo_concurrency.lockutils [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] Acquired lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.418909] env[62952]: DEBUG nova.network.neutron [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Refreshing network info cache for port f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.439146] env[62952]: DEBUG nova.objects.base [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 927.466379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.484534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Releasing lock "refresh_cache-4983f354-77dc-4e84-969b-4c74441fa568" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.484939] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Instance network_info: |[{"id": "1e875b4e-53fe-4d5e-b905-f34fcdcc8e86", "address": "fa:16:3e:89:02:e4", "network": {"id": "b5311ec0-3267-40d6-90ef-119177df8283", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1457467599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "373ca19a1eb64177b43d5e107d192959", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e875b4e-53", "ovs_interfaceid": "1e875b4e-53fe-4d5e-b905-f34fcdcc8e86", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 927.485676] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:02:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e875b4e-53fe-4d5e-b905-f34fcdcc8e86', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.495489] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Creating folder: Project (373ca19a1eb64177b43d5e107d192959). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 927.495922] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5cc9fbd7-3f53-4a4b-b805-bb957752d1f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.510624] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Created folder: Project (373ca19a1eb64177b43d5e107d192959) in parent group-v290852. [ 927.510849] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Creating folder: Instances. Parent ref: group-v290893. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 927.511120] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22783d35-38a9-4ef3-9cad-40244ffcdc9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.522472] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.525045] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Created folder: Instances in parent group-v290893. [ 927.525283] env[62952]: DEBUG oslo.service.loopingcall [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.526075] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 927.530181] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-088bf147-230b-4e2a-8ef3-9fc41215324f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.549421] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.549421] env[62952]: value = "task-1367176" [ 927.549421] env[62952]: _type = "Task" [ 927.549421] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.555091] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.555369] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.555533] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.555714] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.555858] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.556012] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.556299] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.556641] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.556951] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.557215] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.557419] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.558353] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af643db3-a15e-4659-b501-c04931bb150e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.570074] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367176, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.570528] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c83ad3a-425c-48d6-bbb7-a04e02830565 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.784699] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367173, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.826296] env[62952]: DEBUG nova.network.neutron [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.922519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.923163] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 927.930031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.183s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.931615] env[62952]: INFO nova.compute.claims [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.063649] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367176, 'name': CreateVM_Task, 'duration_secs': 0.382702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.064666] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.065560] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.065896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.066818] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.068087] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4568de44-c14d-49c5-96ed-e2887dd8f0ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.080812] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 928.080812] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5276d254-24b6-f825-536d-cdd18ea6820e" [ 928.080812] env[62952]: _type = "Task" [ 928.080812] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.090549] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5276d254-24b6-f825-536d-cdd18ea6820e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.246651] env[62952]: DEBUG nova.network.neutron [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updated VIF entry in instance network info cache for port f80edb6b-5326-4531-a02f-366484e828ce. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.247332] env[62952]: DEBUG nova.network.neutron [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.281884] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655447} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.281884] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 928.281884] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.281884] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37ef0e8d-927a-4cd8-a49b-319ece56fb24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.286818] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 928.286818] env[62952]: value = "task-1367177" [ 928.286818] env[62952]: _type = "Task" [ 928.286818] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.297974] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367177, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.329549] env[62952]: INFO nova.compute.manager [-] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Took 1.18 seconds to deallocate network for instance. [ 928.394578] env[62952]: DEBUG nova.network.neutron [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Updating instance_info_cache with network_info: [{"id": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "address": "fa:16:3e:07:66:c7", "network": {"id": "ef8b562d-d3f7-46ca-8150-049eeccdcab4", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-945343030-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6d1dbb357954f40995c72695f446232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c0d5204b-f60e-4830-84c8-2fe246c28202", "external-id": "nsx-vlan-transportzone-104", "segmentation_id": 104, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea71c6a-30", "ovs_interfaceid": "cea71c6a-30a0-4f56-bc79-6b103c4b59c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.436048] env[62952]: DEBUG nova.compute.utils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.444817] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 928.444817] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 928.536232] env[62952]: DEBUG nova.policy [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '378e043e429e4b3098e0003311d0b6de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f7b88e9cabf41a7802ebd6b3b84d51e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.595401] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5276d254-24b6-f825-536d-cdd18ea6820e, 'name': SearchDatastore_Task, 'duration_secs': 0.011471} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.595750] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.595983] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.596246] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.596391] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.596567] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.596844] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5325e69a-ed90-403b-82e8-f4541881bad3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.605707] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.605904] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.606954] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9c1e445-6a59-4cde-9f96-130e119acf69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.612704] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 928.612704] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52165807-b854-5a6e-f6c5-350d9bc54ef0" [ 928.612704] env[62952]: _type = "Task" [ 928.612704] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.619341] env[62952]: DEBUG nova.compute.manager [req-27435857-5e30-4f26-9eaf-f64c03085b5c req-8acd4d23-27e2-46be-aa70-e3c79ea1529c service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Received event network-vif-plugged-c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.619555] env[62952]: DEBUG oslo_concurrency.lockutils [req-27435857-5e30-4f26-9eaf-f64c03085b5c req-8acd4d23-27e2-46be-aa70-e3c79ea1529c service nova] Acquiring lock "7621d6a6-27aa-45d3-80fa-65e957519a1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.619787] env[62952]: DEBUG oslo_concurrency.lockutils [req-27435857-5e30-4f26-9eaf-f64c03085b5c req-8acd4d23-27e2-46be-aa70-e3c79ea1529c service nova] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.620024] env[62952]: DEBUG oslo_concurrency.lockutils [req-27435857-5e30-4f26-9eaf-f64c03085b5c req-8acd4d23-27e2-46be-aa70-e3c79ea1529c service nova] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.620145] env[62952]: DEBUG nova.compute.manager [req-27435857-5e30-4f26-9eaf-f64c03085b5c req-8acd4d23-27e2-46be-aa70-e3c79ea1529c service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] No waiting events found dispatching network-vif-plugged-c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.620283] env[62952]: WARNING nova.compute.manager [req-27435857-5e30-4f26-9eaf-f64c03085b5c req-8acd4d23-27e2-46be-aa70-e3c79ea1529c service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Received unexpected event network-vif-plugged-c1a7ed7e-0067-4158-95c8-bdc2c8104c1d for instance with vm_state building and task_state spawning. [ 928.624073] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52165807-b854-5a6e-f6c5-350d9bc54ef0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.732976] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Successfully updated port: c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.750150] env[62952]: DEBUG oslo_concurrency.lockutils [req-0d3be3b2-dd5f-477d-a673-58e6769af925 req-77698a8b-3cea-492b-b5dc-4a5a6250ff13 service nova] Releasing lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.798238] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367177, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075185} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.798531] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.799334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e18b50-4fd2-428a-b284-55d3c2f60d0f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.831855] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.832508] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56175469-f95a-407b-8b4c-984fd9dd5fe1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.848759] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.855287] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 928.855287] env[62952]: value = "task-1367178" [ 928.855287] env[62952]: _type = "Task" [ 928.855287] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.865689] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367178, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.897728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Releasing lock "refresh_cache-a63ec2ac-4484-4360-962b-105a5338c1b8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.947610] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.063022] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Successfully created port: fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.123104] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52165807-b854-5a6e-f6c5-350d9bc54ef0, 'name': SearchDatastore_Task, 'duration_secs': 0.008936} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.123951] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81fd313e-71da-4352-9ab2-47125337bccf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.129181] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 929.129181] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52e4fd65-e82c-2bf9-c728-621d2ab078dd" [ 929.129181] env[62952]: _type = "Task" [ 929.129181] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.143621] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52e4fd65-e82c-2bf9-c728-621d2ab078dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.237039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "refresh_cache-7621d6a6-27aa-45d3-80fa-65e957519a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.237197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "refresh_cache-7621d6a6-27aa-45d3-80fa-65e957519a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.237349] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.245188] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6788eb-7c73-4617-9691-3dc3cc1e7a75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.253473] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70a9bff-4f18-4652-9db6-21cf18f54038 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.287363] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25fe30f-5c80-47fd-aa00-6e91e9043e79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.293789] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b62f3e9-81ad-4b3e-8fa4-0d7f084e95c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.307759] env[62952]: DEBUG nova.compute.provider_tree [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.364698] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367178, 'name': ReconfigVM_Task, 'duration_secs': 0.314035} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.364971] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.365669] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35a2be5f-a801-473a-b347-a514dd1bf3e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.373626] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 929.373626] env[62952]: value = "task-1367179" [ 929.373626] env[62952]: _type = "Task" [ 929.373626] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.380219] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367179, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.403962] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.404311] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00e1f091-4c2d-4269-8f2f-259e29227c72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.411710] env[62952]: DEBUG oslo_vmware.api [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 929.411710] env[62952]: value = "task-1367180" [ 929.411710] env[62952]: _type = "Task" [ 929.411710] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.421635] env[62952]: DEBUG oslo_vmware.api [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.639949] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52e4fd65-e82c-2bf9-c728-621d2ab078dd, 'name': SearchDatastore_Task, 'duration_secs': 0.01133} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.640275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.641033] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 4983f354-77dc-4e84-969b-4c74441fa568/4983f354-77dc-4e84-969b-4c74441fa568.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.641033] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08caeac2-3645-4009-8883-63b1cdea7080 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.647858] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 929.647858] env[62952]: value = "task-1367181" [ 929.647858] env[62952]: _type = "Task" [ 929.647858] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.656450] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.778042] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 929.811045] env[62952]: DEBUG nova.scheduler.client.report [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.883834] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367179, 'name': Rename_Task, 'duration_secs': 0.147251} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.884567] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.884878] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d448cb7-f0a7-4c57-99e7-7bbb41b78f73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.891914] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 929.891914] env[62952]: value = "task-1367182" [ 929.891914] env[62952]: _type = "Task" [ 929.891914] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.908190] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.923632] env[62952]: DEBUG oslo_vmware.api [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367180, 'name': PowerOnVM_Task, 'duration_secs': 0.409695} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.923632] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.923632] env[62952]: DEBUG nova.compute.manager [None req-0a583833-075a-4847-ae44-0486721273df tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.924086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559b412a-b6cc-4b96-86c4-cc067e86ebdf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.958394] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.004415] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.004809] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.005248] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.005248] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.005248] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.005365] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.005722] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.005801] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.005920] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.006214] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.006348] env[62952]: DEBUG nova.virt.hardware [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.007643] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8b874b-9f5d-4b10-88ce-6c8e190f6dc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.016687] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ccdf50-89a3-4bcd-bde5-f00b93d5ebe8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.022699] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Updating instance_info_cache with network_info: [{"id": "c1a7ed7e-0067-4158-95c8-bdc2c8104c1d", "address": "fa:16:3e:d6:08:8f", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1a7ed7e-00", "ovs_interfaceid": "c1a7ed7e-0067-4158-95c8-bdc2c8104c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.160074] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367181, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.318289] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.318836] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 930.323257] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.022s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.325237] env[62952]: INFO nova.compute.claims [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.403262] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367182, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.526039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "refresh_cache-7621d6a6-27aa-45d3-80fa-65e957519a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.526382] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Instance network_info: |[{"id": "c1a7ed7e-0067-4158-95c8-bdc2c8104c1d", "address": "fa:16:3e:d6:08:8f", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1a7ed7e-00", "ovs_interfaceid": "c1a7ed7e-0067-4158-95c8-bdc2c8104c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.526804] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:08:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c1a7ed7e-0067-4158-95c8-bdc2c8104c1d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.535046] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Creating folder: Project (8f7b88e9cabf41a7802ebd6b3b84d51e). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.539378] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa3c3e05-2516-40ae-8983-5d29d7366371 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.553152] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Created folder: Project (8f7b88e9cabf41a7802ebd6b3b84d51e) in parent group-v290852. [ 930.553454] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Creating folder: Instances. Parent ref: group-v290896. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.555133] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12564a93-3a96-4639-b836-474fa3fd565f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.564436] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Created folder: Instances in parent group-v290896. [ 930.564893] env[62952]: DEBUG oslo.service.loopingcall [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.569019] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.569019] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26ecab3b-53c1-4136-aa49-3c9b0a8da711 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.593498] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.593498] env[62952]: value = "task-1367185" [ 930.593498] env[62952]: _type = "Task" [ 930.593498] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.603751] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367185, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.659036] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367181, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513265} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.659367] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 4983f354-77dc-4e84-969b-4c74441fa568/4983f354-77dc-4e84-969b-4c74441fa568.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.659766] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.660286] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e68dc360-d136-4fd2-9b5b-7b5637029750 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.667378] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 930.667378] env[62952]: value = "task-1367186" [ 930.667378] env[62952]: _type = "Task" [ 930.667378] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.675237] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.700550] env[62952]: DEBUG nova.compute.manager [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Received event network-changed-c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.700754] env[62952]: DEBUG nova.compute.manager [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Refreshing instance network info cache due to event network-changed-c1a7ed7e-0067-4158-95c8-bdc2c8104c1d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.700955] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] Acquiring lock "refresh_cache-7621d6a6-27aa-45d3-80fa-65e957519a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.701093] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] Acquired lock "refresh_cache-7621d6a6-27aa-45d3-80fa-65e957519a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.701268] env[62952]: DEBUG nova.network.neutron [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Refreshing network info cache for port c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.824880] env[62952]: DEBUG nova.compute.utils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.826647] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.826896] env[62952]: DEBUG nova.network.neutron [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.867786] env[62952]: DEBUG nova.policy [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55ffb58a3b3a4ad29699d48ecd6c0700', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d564dc3c75e430dbaf1f90a9c90c18b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.903226] env[62952]: DEBUG oslo_vmware.api [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367182, 'name': PowerOnVM_Task, 'duration_secs': 0.775197} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.903356] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.903619] env[62952]: INFO nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Took 10.07 seconds to spawn the instance on the hypervisor. [ 930.904010] env[62952]: DEBUG nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.904802] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddeb80a-91c7-42d1-b60b-588d900977ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.102585] env[62952]: DEBUG nova.compute.manager [req-59a5b579-7c7c-4f1c-8c22-8a69edbfa207 req-796c4f82-2027-409c-8076-204cca515963 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Received event network-vif-plugged-fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.102811] env[62952]: DEBUG oslo_concurrency.lockutils [req-59a5b579-7c7c-4f1c-8c22-8a69edbfa207 req-796c4f82-2027-409c-8076-204cca515963 service nova] Acquiring lock "93951b09-9d7c-42da-83de-0db3980d0d5d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.103039] env[62952]: DEBUG oslo_concurrency.lockutils [req-59a5b579-7c7c-4f1c-8c22-8a69edbfa207 req-796c4f82-2027-409c-8076-204cca515963 service nova] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.103215] env[62952]: DEBUG oslo_concurrency.lockutils [req-59a5b579-7c7c-4f1c-8c22-8a69edbfa207 req-796c4f82-2027-409c-8076-204cca515963 service nova] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.103383] env[62952]: DEBUG nova.compute.manager [req-59a5b579-7c7c-4f1c-8c22-8a69edbfa207 req-796c4f82-2027-409c-8076-204cca515963 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] No waiting events found dispatching network-vif-plugged-fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.103583] env[62952]: WARNING nova.compute.manager [req-59a5b579-7c7c-4f1c-8c22-8a69edbfa207 req-796c4f82-2027-409c-8076-204cca515963 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Received unexpected event network-vif-plugged-fd33b4df-02a6-48ec-8118-6ffc34a8e633 for instance with vm_state building and task_state spawning. [ 931.109750] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367185, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.185719] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101826} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.186129] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.187272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8562781e-2727-453b-ad82-6ebc98f4bf2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.227023] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 4983f354-77dc-4e84-969b-4c74441fa568/4983f354-77dc-4e84-969b-4c74441fa568.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.227023] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efe776fa-d8eb-49e9-b733-6902f21cc859 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.245100] env[62952]: DEBUG nova.network.neutron [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Successfully created port: b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 931.255018] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 931.255018] env[62952]: value = "task-1367187" [ 931.255018] env[62952]: _type = "Task" [ 931.255018] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.264136] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367187, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.308815] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Successfully updated port: fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.329654] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 931.435018] env[62952]: INFO nova.compute.manager [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Took 32.52 seconds to build instance. [ 931.608533] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367185, 'name': CreateVM_Task, 'duration_secs': 0.836456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.608705] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.609384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.609541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.609941] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.610628] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c41c83a6-2d67-4f49-9b23-7d2a17d05f5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.615834] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 931.615834] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bfc2d1-2198-b173-ea78-1cfce80223e6" [ 931.615834] env[62952]: _type = "Task" [ 931.615834] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.629259] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52bfc2d1-2198-b173-ea78-1cfce80223e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.654237] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07dd4fbc-88cb-403f-a09e-f6cee219ec5c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.663089] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08f8726-00c4-478c-b741-fd43940fab51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.698831] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca34866f-4101-46c1-9072-2dd08da05f03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.706579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-124f02fd-a809-4c45-acf7-f2b8bcd4570b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.724936] env[62952]: DEBUG nova.compute.provider_tree [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.765043] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367187, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.810956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "refresh_cache-93951b09-9d7c-42da-83de-0db3980d0d5d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.810956] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "refresh_cache-93951b09-9d7c-42da-83de-0db3980d0d5d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.811190] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.827041] env[62952]: DEBUG nova.network.neutron [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Updated VIF entry in instance network info cache for port c1a7ed7e-0067-4158-95c8-bdc2c8104c1d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.827041] env[62952]: DEBUG nova.network.neutron [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Updating instance_info_cache with network_info: [{"id": "c1a7ed7e-0067-4158-95c8-bdc2c8104c1d", "address": "fa:16:3e:d6:08:8f", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1a7ed7e-00", "ovs_interfaceid": "c1a7ed7e-0067-4158-95c8-bdc2c8104c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.937686] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3910890f-daf8-43cb-ba50-6762b3917a65 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.837s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.127232] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52bfc2d1-2198-b173-ea78-1cfce80223e6, 'name': SearchDatastore_Task, 'duration_secs': 0.013015} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.127602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.127780] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.128030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.128181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.128357] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.128629] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80a6801d-b4b7-48f4-b43c-85e7e1e369ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.146807] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.146969] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.147754] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0011a6e3-9546-410d-8e12-69f91bd09582 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.155329] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 932.155329] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ae7ee4-174f-4387-e944-e2a2ff40d467" [ 932.155329] env[62952]: _type = "Task" [ 932.155329] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.163396] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ae7ee4-174f-4387-e944-e2a2ff40d467, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.228237] env[62952]: DEBUG nova.scheduler.client.report [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.270106] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367187, 'name': ReconfigVM_Task, 'duration_secs': 0.520371} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.270106] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 4983f354-77dc-4e84-969b-4c74441fa568/4983f354-77dc-4e84-969b-4c74441fa568.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.270106] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51f497c3-9031-4d20-9ba8-4997845d216a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.278080] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 932.278080] env[62952]: value = "task-1367188" [ 932.278080] env[62952]: _type = "Task" [ 932.278080] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.287578] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367188, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.329420] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a14d9ae-d2b9-49a9-ad4c-75170c6ef004 req-96d76578-1a2c-4169-85e8-e30379d1346d service nova] Releasing lock "refresh_cache-7621d6a6-27aa-45d3-80fa-65e957519a1f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.338268] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 932.359984] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.360271] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.360432] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.360617] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.360836] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.360904] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.361847] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.361917] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.362081] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.362257] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.362461] env[62952]: DEBUG nova.virt.hardware [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.363338] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1486ae29-d865-4ddc-9576-d8f7a8fc4aeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.371618] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c20291-dde0-43df-844f-e58e0f9f112a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.388285] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.440393] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.511413] env[62952]: INFO nova.compute.manager [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Rebuilding instance [ 932.568650] env[62952]: DEBUG nova.compute.manager [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.569564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a94a5f7-f500-4b32-b66b-653a419baea0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.669359] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ae7ee4-174f-4387-e944-e2a2ff40d467, 'name': SearchDatastore_Task, 'duration_secs': 0.047078} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.671271] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4acf81c-8f09-444f-ae04-be8877c2852d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.676801] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 932.676801] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52907d48-cce8-a46f-405e-42a4b33496fd" [ 932.676801] env[62952]: _type = "Task" [ 932.676801] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.686594] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52907d48-cce8-a46f-405e-42a4b33496fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.734725] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.735297] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.739047] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.385s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.739218] env[62952]: DEBUG nova.objects.instance [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lazy-loading 'resources' on Instance uuid d2faf1e0-cfd1-4d87-ba77-0af92dc16643 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.790339] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367188, 'name': Rename_Task, 'duration_secs': 0.47594} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.790960] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.791241] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-383fc629-9722-470e-8609-c05c6dd4237b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.803110] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 932.803110] env[62952]: value = "task-1367189" [ 932.803110] env[62952]: _type = "Task" [ 932.803110] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.813374] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.878057] env[62952]: DEBUG nova.network.neutron [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Updating instance_info_cache with network_info: [{"id": "fd33b4df-02a6-48ec-8118-6ffc34a8e633", "address": "fa:16:3e:8b:08:7e", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd33b4df-02", "ovs_interfaceid": "fd33b4df-02a6-48ec-8118-6ffc34a8e633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.975969] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.084208] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.084208] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a462d60-50ea-4b0d-be1d-b93801874fce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.090117] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 933.090117] env[62952]: value = "task-1367190" [ 933.090117] env[62952]: _type = "Task" [ 933.090117] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.101352] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.145270] env[62952]: DEBUG nova.compute.manager [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Received event network-changed-fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.145410] env[62952]: DEBUG nova.compute.manager [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Refreshing instance network info cache due to event network-changed-fd33b4df-02a6-48ec-8118-6ffc34a8e633. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.145644] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] Acquiring lock "refresh_cache-93951b09-9d7c-42da-83de-0db3980d0d5d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.189016] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52907d48-cce8-a46f-405e-42a4b33496fd, 'name': SearchDatastore_Task, 'duration_secs': 0.011685} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.189318] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.189591] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 7621d6a6-27aa-45d3-80fa-65e957519a1f/7621d6a6-27aa-45d3-80fa-65e957519a1f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.189861] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ad8f4db-5fbe-4067-9b38-7e4c6957e6f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.197838] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 933.197838] env[62952]: value = "task-1367191" [ 933.197838] env[62952]: _type = "Task" [ 933.197838] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.205848] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367191, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.242467] env[62952]: DEBUG nova.compute.utils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.251930] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.252232] env[62952]: DEBUG nova.network.neutron [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.317542] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367189, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.364024] env[62952]: DEBUG nova.policy [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b4bcad9f5ed4ee7882df428957d1d93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8427597d4aae451e9375ade7c06b2bcf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.381026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "refresh_cache-93951b09-9d7c-42da-83de-0db3980d0d5d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.381026] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Instance network_info: |[{"id": "fd33b4df-02a6-48ec-8118-6ffc34a8e633", "address": "fa:16:3e:8b:08:7e", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd33b4df-02", "ovs_interfaceid": "fd33b4df-02a6-48ec-8118-6ffc34a8e633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.381457] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] Acquired lock "refresh_cache-93951b09-9d7c-42da-83de-0db3980d0d5d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.381457] env[62952]: DEBUG nova.network.neutron [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Refreshing network info cache for port fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.382435] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:08:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd33b4df-02a6-48ec-8118-6ffc34a8e633', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.391085] env[62952]: DEBUG oslo.service.loopingcall [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.396868] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.398156] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bbed1001-0b18-40ed-b9ed-307e98868dc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.419663] env[62952]: DEBUG nova.compute.manager [req-6d22b87a-3d24-4067-983c-aa2c24404e7f req-32f0ec07-cfbb-4f70-88fa-bb8d1656ecd3 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Received event network-vif-plugged-b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.420030] env[62952]: DEBUG oslo_concurrency.lockutils [req-6d22b87a-3d24-4067-983c-aa2c24404e7f req-32f0ec07-cfbb-4f70-88fa-bb8d1656ecd3 service nova] Acquiring lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.420331] env[62952]: DEBUG oslo_concurrency.lockutils [req-6d22b87a-3d24-4067-983c-aa2c24404e7f req-32f0ec07-cfbb-4f70-88fa-bb8d1656ecd3 service nova] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.420496] env[62952]: DEBUG oslo_concurrency.lockutils [req-6d22b87a-3d24-4067-983c-aa2c24404e7f req-32f0ec07-cfbb-4f70-88fa-bb8d1656ecd3 service nova] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.420666] env[62952]: DEBUG nova.compute.manager [req-6d22b87a-3d24-4067-983c-aa2c24404e7f req-32f0ec07-cfbb-4f70-88fa-bb8d1656ecd3 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] No waiting events found dispatching network-vif-plugged-b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.420834] env[62952]: WARNING nova.compute.manager [req-6d22b87a-3d24-4067-983c-aa2c24404e7f req-32f0ec07-cfbb-4f70-88fa-bb8d1656ecd3 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Received unexpected event network-vif-plugged-b0c7b851-c731-4055-b489-db9ecae60444 for instance with vm_state building and task_state spawning. [ 933.430564] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.430564] env[62952]: value = "task-1367192" [ 933.430564] env[62952]: _type = "Task" [ 933.430564] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.438161] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367192, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.598571] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3ed701-4f61-40cf-8a75-9b8c56305a00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.611903] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.613986] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228c086a-3388-453a-ac1a-36974d5636e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.650788] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf76d15d-2aa4-47fb-ac0e-3abd041c79bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.659502] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b7cbc5-e948-4fa7-98ea-21c2f1cbc11f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.674156] env[62952]: DEBUG nova.compute.provider_tree [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.705503] env[62952]: DEBUG nova.network.neutron [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Updated VIF entry in instance network info cache for port fd33b4df-02a6-48ec-8118-6ffc34a8e633. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.705503] env[62952]: DEBUG nova.network.neutron [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Updating instance_info_cache with network_info: [{"id": "fd33b4df-02a6-48ec-8118-6ffc34a8e633", "address": "fa:16:3e:8b:08:7e", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd33b4df-02", "ovs_interfaceid": "fd33b4df-02a6-48ec-8118-6ffc34a8e633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.705853] env[62952]: DEBUG nova.network.neutron [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Successfully updated port: b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.715499] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367191, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.751618] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.820049] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367189, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.924978] env[62952]: DEBUG nova.network.neutron [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Successfully created port: aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.938405] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367192, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.107206] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.107730] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367190, 'name': PowerOffVM_Task, 'duration_secs': 0.800784} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.107987] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.109789] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.110255] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.113647] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823f5445-38ba-4020-b8e5-86dc01768fa6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.121223] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.121384] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5aa40823-cfbb-4980-a330-af7d4c7393f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.177245] env[62952]: DEBUG nova.scheduler.client.report [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.205743] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.206610] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.206610] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.206952] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ec0cff1-157b-409f-84ce-23348fc4e2e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.214840] env[62952]: DEBUG oslo_concurrency.lockutils [req-dc72d37e-c1c6-438d-89ae-f93c05770928 req-335a3df2-b91e-4a71-a090-80555a2565e2 service nova] Releasing lock "refresh_cache-93951b09-9d7c-42da-83de-0db3980d0d5d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.214840] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "refresh_cache-94fd5287-9a8e-45e4-99e8-0bcc861f889c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.214966] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "refresh_cache-94fd5287-9a8e-45e4-99e8-0bcc861f889c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.215095] env[62952]: DEBUG nova.network.neutron [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.216734] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367191, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5375} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.217515] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 7621d6a6-27aa-45d3-80fa-65e957519a1f/7621d6a6-27aa-45d3-80fa-65e957519a1f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.217763] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.219082] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b63fe416-9734-4fbc-a64c-2c4058bd3ee9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.222095] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 934.222095] env[62952]: value = "task-1367194" [ 934.222095] env[62952]: _type = "Task" [ 934.222095] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.231380] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 934.231380] env[62952]: value = "task-1367195" [ 934.231380] env[62952]: _type = "Task" [ 934.231380] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.239730] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.244968] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367195, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.319762] env[62952]: DEBUG oslo_vmware.api [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367189, 'name': PowerOnVM_Task, 'duration_secs': 1.426494} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.320152] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.320435] env[62952]: INFO nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Took 11.16 seconds to spawn the instance on the hypervisor. [ 934.320673] env[62952]: DEBUG nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.321558] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8613242a-92b0-4ead-98d4-3a4fad2e2118 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.439084] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367192, 'name': CreateVM_Task, 'duration_secs': 0.748594} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.439269] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.439954] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.440137] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.440474] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.440701] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cbdaff9-aaae-4bc2-ad4f-78794da7d321 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.445743] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 934.445743] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]525f1e0a-a534-6982-ed1a-caa596726864" [ 934.445743] env[62952]: _type = "Task" [ 934.445743] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.453960] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525f1e0a-a534-6982-ed1a-caa596726864, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.620147] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.620316] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 934.620432] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 934.671036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "2f475040-28df-47e4-bfc4-4c843a57885f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.671036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "2f475040-28df-47e4-bfc4-4c843a57885f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.671036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "2f475040-28df-47e4-bfc4-4c843a57885f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.671036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "2f475040-28df-47e4-bfc4-4c843a57885f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.671271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "2f475040-28df-47e4-bfc4-4c843a57885f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.673879] env[62952]: INFO nova.compute.manager [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Terminating instance [ 934.675861] env[62952]: DEBUG nova.compute.manager [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.676078] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.676905] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272849fa-449d-4e22-8486-93573a500494 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.682590] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.687932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.118s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.688654] env[62952]: INFO nova.compute.claims [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.695015] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.695015] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf8e9655-4ef5-40e5-a8b6-edd488421583 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.701622] env[62952]: DEBUG oslo_vmware.api [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 934.701622] env[62952]: value = "task-1367196" [ 934.701622] env[62952]: _type = "Task" [ 934.701622] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.710447] env[62952]: DEBUG oslo_vmware.api [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367196, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.711761] env[62952]: INFO nova.scheduler.client.report [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Deleted allocations for instance d2faf1e0-cfd1-4d87-ba77-0af92dc16643 [ 934.734118] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.744848] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367195, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.189312} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.744848] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.744848] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f67b31-dd3b-460b-9238-351df0bce32e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.768828] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 7621d6a6-27aa-45d3-80fa-65e957519a1f/7621d6a6-27aa-45d3-80fa-65e957519a1f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.769898] env[62952]: DEBUG nova.network.neutron [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 934.773349] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.776158] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4e827dc-2e07-4c9d-b95f-bf5d8ee075e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.798998] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 934.798998] env[62952]: value = "task-1367197" [ 934.798998] env[62952]: _type = "Task" [ 934.798998] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.808370] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367197, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.822478] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.822844] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.822980] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.823441] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.823441] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.823570] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.823944] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.824076] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.824255] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.824419] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.824589] env[62952]: DEBUG nova.virt.hardware [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.825718] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6aea502-d3e2-44a1-aa8f-ee63ae5b5977 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.833417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba72620-2a6c-4a04-8e5f-918bed22d0dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.854247] env[62952]: INFO nova.compute.manager [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Took 28.97 seconds to build instance. [ 934.868030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.868266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.955846] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525f1e0a-a534-6982-ed1a-caa596726864, 'name': SearchDatastore_Task, 'duration_secs': 0.026633} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.956356] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.956606] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.956845] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.956991] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.957387] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.957642] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e32e8e9-cde4-4521-961c-427e28c21679 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.967081] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.967260] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 934.967940] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45d43149-04fb-4ac1-8224-6b1cac1154d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.973192] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 934.973192] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523d4e77-939f-30a7-50b6-6599490fca22" [ 934.973192] env[62952]: _type = "Task" [ 934.973192] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.982709] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523d4e77-939f-30a7-50b6-6599490fca22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.035493] env[62952]: DEBUG nova.network.neutron [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Updating instance_info_cache with network_info: [{"id": "b0c7b851-c731-4055-b489-db9ecae60444", "address": "fa:16:3e:44:df:c9", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0c7b851-c7", "ovs_interfaceid": "b0c7b851-c731-4055-b489-db9ecae60444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.126619] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Skipping network cache update for instance because it is being deleted. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 935.126792] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 935.126927] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 935.127084] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 935.127215] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Skipping network cache update for instance because it is Building. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 935.146098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.146339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.146690] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 935.146690] env[62952]: DEBUG nova.objects.instance [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lazy-loading 'info_cache' on Instance uuid d2faf1e0-cfd1-4d87-ba77-0af92dc16643 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.217027] env[62952]: DEBUG oslo_vmware.api [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367196, 'name': PowerOffVM_Task, 'duration_secs': 0.201294} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.217515] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.217515] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.217718] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c327c78c-479b-4955-bafd-5278a6869fc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.231615] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb96172-b588-454c-9b8a-4f03008ded2d tempest-ServerAddressesNegativeTestJSON-17476672 tempest-ServerAddressesNegativeTestJSON-17476672-project-member] Lock "d2faf1e0-cfd1-4d87-ba77-0af92dc16643" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.812s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.239546] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.521961} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.239718] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.239902] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.240092] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.309636] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367197, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.328961] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.329133] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.330013] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleting the datastore file [datastore2] 2f475040-28df-47e4-bfc4-4c843a57885f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.330013] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b33a1b54-7de9-4e72-8cc9-b938bab19848 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.336253] env[62952]: DEBUG oslo_vmware.api [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 935.336253] env[62952]: value = "task-1367199" [ 935.336253] env[62952]: _type = "Task" [ 935.336253] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.344567] env[62952]: DEBUG oslo_vmware.api [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367199, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.357212] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ed8b6b-413e-4926-803c-ed31e7425389 tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "4983f354-77dc-4e84-969b-4c74441fa568" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.226s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.484900] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523d4e77-939f-30a7-50b6-6599490fca22, 'name': SearchDatastore_Task, 'duration_secs': 0.024746} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.486331] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a505c279-47de-4167-8338-6d6c76d0907e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.492525] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 935.492525] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5270fdff-df97-21a5-478d-0a9724db3faa" [ 935.492525] env[62952]: _type = "Task" [ 935.492525] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.503251] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5270fdff-df97-21a5-478d-0a9724db3faa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.538057] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "refresh_cache-94fd5287-9a8e-45e4-99e8-0bcc861f889c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.538392] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Instance network_info: |[{"id": "b0c7b851-c731-4055-b489-db9ecae60444", "address": "fa:16:3e:44:df:c9", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0c7b851-c7", "ovs_interfaceid": "b0c7b851-c731-4055-b489-db9ecae60444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 935.538804] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:df:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92fe29b3-0907-453d-aabb-5559c4bd7c0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0c7b851-c731-4055-b489-db9ecae60444', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.550449] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating folder: Project (6d564dc3c75e430dbaf1f90a9c90c18b). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 935.550750] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fedc95d6-fb42-45da-b288-fdaaa96b8645 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.559578] env[62952]: DEBUG nova.compute.manager [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Received event network-changed-b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.559758] env[62952]: DEBUG nova.compute.manager [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Refreshing instance network info cache due to event network-changed-b0c7b851-c731-4055-b489-db9ecae60444. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.561849] env[62952]: DEBUG oslo_concurrency.lockutils [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] Acquiring lock "refresh_cache-94fd5287-9a8e-45e4-99e8-0bcc861f889c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.561849] env[62952]: DEBUG oslo_concurrency.lockutils [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] Acquired lock "refresh_cache-94fd5287-9a8e-45e4-99e8-0bcc861f889c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.561849] env[62952]: DEBUG nova.network.neutron [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Refreshing network info cache for port b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.562784] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created folder: Project (6d564dc3c75e430dbaf1f90a9c90c18b) in parent group-v290852. [ 935.563123] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating folder: Instances. Parent ref: group-v290900. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 935.563459] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40e5f44a-5d29-4d11-9050-37e4c4a50240 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.575887] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created folder: Instances in parent group-v290900. [ 935.576203] env[62952]: DEBUG oslo.service.loopingcall [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.576428] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.576661] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98087f9f-ae27-4fcf-bdc5-410df4e67b62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.602235] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.602235] env[62952]: value = "task-1367202" [ 935.602235] env[62952]: _type = "Task" [ 935.602235] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.613263] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367202, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.617907] env[62952]: DEBUG nova.network.neutron [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Successfully updated port: aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.811261] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367197, 'name': ReconfigVM_Task, 'duration_secs': 0.897991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.811507] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 7621d6a6-27aa-45d3-80fa-65e957519a1f/7621d6a6-27aa-45d3-80fa-65e957519a1f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.812176] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-322360e3-755b-461f-855a-f5ef1f6f29fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.818530] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 935.818530] env[62952]: value = "task-1367203" [ 935.818530] env[62952]: _type = "Task" [ 935.818530] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.832354] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367203, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.847328] env[62952]: DEBUG oslo_vmware.api [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367199, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.366199} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.849880] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.850065] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.850247] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.850492] env[62952]: INFO nova.compute.manager [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Took 1.17 seconds to destroy the instance on the hypervisor. [ 935.850646] env[62952]: DEBUG oslo.service.loopingcall [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.851026] env[62952]: DEBUG nova.compute.manager [-] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.851163] env[62952]: DEBUG nova.network.neutron [-] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.859546] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.004747] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5270fdff-df97-21a5-478d-0a9724db3faa, 'name': SearchDatastore_Task, 'duration_secs': 0.022606} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.005096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.005361] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 93951b09-9d7c-42da-83de-0db3980d0d5d/93951b09-9d7c-42da-83de-0db3980d0d5d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.005801] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c369e42c-c51d-4453-bd7c-72431894aac1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.020229] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 936.020229] env[62952]: value = "task-1367204" [ 936.020229] env[62952]: _type = "Task" [ 936.020229] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.029762] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.031861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212d0c5c-3dd8-4687-bf17-96e7feef8692 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.038759] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7706e7-e318-468c-944b-89c19bde1fbd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.070536] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab2c276-d380-4121-b412-b83cb3e845d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.078209] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76233500-3870-4b55-9487-c669c9dab9e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.095355] env[62952]: DEBUG nova.compute.provider_tree [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.113742] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367202, 'name': CreateVM_Task, 'duration_secs': 0.459627} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.114109] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.114582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.114770] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.115108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.115359] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eee01926-969a-4b7f-85e4-d4ee329e033a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.120036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.120398] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.120398] env[62952]: DEBUG nova.network.neutron [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.124903] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 936.124903] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526ae9e8-ec92-ca92-94eb-fb356a01a47e" [ 936.124903] env[62952]: _type = "Task" [ 936.124903] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.132115] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526ae9e8-ec92-ca92-94eb-fb356a01a47e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.153010] env[62952]: DEBUG nova.compute.utils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Can not refresh info_cache because instance was not found {{(pid=62952) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 936.173145] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.283282] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.283535] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.283688] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.283907] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.284284] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.284462] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.284667] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.284844] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.285017] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.285214] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.285397] env[62952]: DEBUG nova.virt.hardware [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.286776] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8447bde0-f443-46d4-990a-5ef7d8a298b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.300658] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f9afa8-4b1d-4dd1-8b78-a96056674211 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.321497] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:38:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'feeba8e0-a5e3-4e76-9991-96695e8032eb', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.330654] env[62952]: DEBUG oslo.service.loopingcall [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.331211] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.334546] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0e3e050-5bdd-4c08-b2f8-2f8a59d34285 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.361528] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367203, 'name': Rename_Task, 'duration_secs': 0.1711} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.362916] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.363196] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.363196] env[62952]: value = "task-1367205" [ 936.363196] env[62952]: _type = "Task" [ 936.363196] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.363396] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40dae053-b595-4a08-b931-7c2cf9d66a08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.378477] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367205, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.378477] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 936.378477] env[62952]: value = "task-1367206" [ 936.378477] env[62952]: _type = "Task" [ 936.378477] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.386928] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.395291] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.531874] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367204, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.589929] env[62952]: DEBUG nova.network.neutron [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Updated VIF entry in instance network info cache for port b0c7b851-c731-4055-b489-db9ecae60444. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 936.589929] env[62952]: DEBUG nova.network.neutron [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Updating instance_info_cache with network_info: [{"id": "b0c7b851-c731-4055-b489-db9ecae60444", "address": "fa:16:3e:44:df:c9", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0c7b851-c7", "ovs_interfaceid": "b0c7b851-c731-4055-b489-db9ecae60444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.598607] env[62952]: DEBUG nova.scheduler.client.report [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.641280] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526ae9e8-ec92-ca92-94eb-fb356a01a47e, 'name': SearchDatastore_Task, 'duration_secs': 0.012407} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.641606] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.642369] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.642932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.642932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.643113] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.644632] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f3f8b46-90f9-4e87-9a48-a5ce8d547574 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.656395] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.656554] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.657331] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6a2739c-a865-4d1d-aee8-ea15829dd38a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.664713] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 936.664713] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]527f2a5a-785d-85b8-6c7f-d077e5e3064c" [ 936.664713] env[62952]: _type = "Task" [ 936.664713] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.679356] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527f2a5a-785d-85b8-6c7f-d077e5e3064c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.695933] env[62952]: DEBUG nova.network.neutron [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.878324] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367205, 'name': CreateVM_Task, 'duration_secs': 0.383754} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.878556] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.879313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.879490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.879842] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.883976] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a28a0287-1e64-43bb-832a-42d435c6d715 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.892165] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.893823] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 936.893823] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523ff4d8-1958-27a2-d101-2867109a8308" [ 936.893823] env[62952]: _type = "Task" [ 936.893823] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.905282] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523ff4d8-1958-27a2-d101-2867109a8308, 'name': SearchDatastore_Task, 'duration_secs': 0.011618} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.905629] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.906286] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.906674] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.906827] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.908405] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.908978] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-155ce345-b96d-4474-b0bc-488469d0aabf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.919025] env[62952]: DEBUG nova.network.neutron [-] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.921220] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.921468] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.922248] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3d0be51-05a1-480d-a501-f65fceacba40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.927848] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 936.927848] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52def0cc-9f86-0fe3-3e32-9ccba690d998" [ 936.927848] env[62952]: _type = "Task" [ 936.927848] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.938854] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52def0cc-9f86-0fe3-3e32-9ccba690d998, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.031714] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367204, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701148} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.032063] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 93951b09-9d7c-42da-83de-0db3980d0d5d/93951b09-9d7c-42da-83de-0db3980d0d5d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.032314] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.032584] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4ecfd38-2dbd-4774-a65e-b29a83749227 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.040618] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 937.040618] env[62952]: value = "task-1367207" [ 937.040618] env[62952]: _type = "Task" [ 937.040618] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.051110] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367207, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.092862] env[62952]: DEBUG oslo_concurrency.lockutils [req-ed3fa8ad-f756-4aa7-9227-753a527ff550 req-1faaa813-7732-43e6-b2bf-e7b939e477e4 service nova] Releasing lock "refresh_cache-94fd5287-9a8e-45e4-99e8-0bcc861f889c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.107905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.108664] env[62952]: DEBUG nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.116039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.628s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.116039] env[62952]: INFO nova.compute.claims [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.170209] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.178772] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527f2a5a-785d-85b8-6c7f-d077e5e3064c, 'name': SearchDatastore_Task, 'duration_secs': 0.040498} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.180010] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67060c12-a8e8-4be7-b02b-adc7536b4b7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.186367] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 937.186367] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]529ae3da-2359-c1b3-b1d5-fa91ebd28e83" [ 937.186367] env[62952]: _type = "Task" [ 937.186367] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.194886] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529ae3da-2359-c1b3-b1d5-fa91ebd28e83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.280011] env[62952]: DEBUG nova.network.neutron [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [{"id": "aa4656a7-42d8-47ba-a29d-817000a4b596", "address": "fa:16:3e:14:80:1d", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa4656a7-42", "ovs_interfaceid": "aa4656a7-42d8-47ba-a29d-817000a4b596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.392237] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367206, 'name': PowerOnVM_Task, 'duration_secs': 0.747751} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.392571] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.392748] env[62952]: INFO nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Took 9.87 seconds to spawn the instance on the hypervisor. [ 937.392983] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.393842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d45e7d-04af-4ba9-9053-95972f5123f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.423340] env[62952]: INFO nova.compute.manager [-] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Took 1.57 seconds to deallocate network for instance. [ 937.441027] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52def0cc-9f86-0fe3-3e32-9ccba690d998, 'name': SearchDatastore_Task, 'duration_secs': 0.008671} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.442145] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7228f89-b41e-4972-96a5-e2e8a49e470d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.447774] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 937.447774] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]520c4a7b-1498-f25e-0a91-b5f3156a502a" [ 937.447774] env[62952]: _type = "Task" [ 937.447774] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.457304] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520c4a7b-1498-f25e-0a91-b5f3156a502a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.553212] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367207, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091679} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.553536] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.554363] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96addc6-a3e9-4abb-9bc6-a0ba509f5d69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.581105] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 93951b09-9d7c-42da-83de-0db3980d0d5d/93951b09-9d7c-42da-83de-0db3980d0d5d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.581424] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d866ca05-b2f4-4d97-98e8-e04d337dde43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.599124] env[62952]: DEBUG nova.compute.manager [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Received event network-vif-plugged-aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.599352] env[62952]: DEBUG oslo_concurrency.lockutils [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.599792] env[62952]: DEBUG oslo_concurrency.lockutils [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.599970] env[62952]: DEBUG oslo_concurrency.lockutils [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.600151] env[62952]: DEBUG nova.compute.manager [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] No waiting events found dispatching network-vif-plugged-aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.600317] env[62952]: WARNING nova.compute.manager [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Received unexpected event network-vif-plugged-aa4656a7-42d8-47ba-a29d-817000a4b596 for instance with vm_state building and task_state spawning. [ 937.600474] env[62952]: DEBUG nova.compute.manager [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Received event network-changed-aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.600624] env[62952]: DEBUG nova.compute.manager [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Refreshing instance network info cache due to event network-changed-aa4656a7-42d8-47ba-a29d-817000a4b596. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.600785] env[62952]: DEBUG oslo_concurrency.lockutils [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] Acquiring lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.607457] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 937.607457] env[62952]: value = "task-1367208" [ 937.607457] env[62952]: _type = "Task" [ 937.607457] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.614501] env[62952]: DEBUG nova.compute.utils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.616344] env[62952]: DEBUG nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 937.622712] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367208, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.672718] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-d2faf1e0-cfd1-4d87-ba77-0af92dc16643" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.672907] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 937.673131] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.673299] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.673578] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.673578] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.673729] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.673904] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.674052] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 937.674208] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.696929] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529ae3da-2359-c1b3-b1d5-fa91ebd28e83, 'name': SearchDatastore_Task, 'duration_secs': 0.061505} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.697243] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.697491] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 94fd5287-9a8e-45e4-99e8-0bcc861f889c/94fd5287-9a8e-45e4-99e8-0bcc861f889c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.697736] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-704f3420-8a9b-47bf-9750-6821139e0d00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.704769] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 937.704769] env[62952]: value = "task-1367209" [ 937.704769] env[62952]: _type = "Task" [ 937.704769] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.712574] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367209, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.762852] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "4983f354-77dc-4e84-969b-4c74441fa568" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.763139] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "4983f354-77dc-4e84-969b-4c74441fa568" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.763346] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "4983f354-77dc-4e84-969b-4c74441fa568-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.763531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "4983f354-77dc-4e84-969b-4c74441fa568-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.763701] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "4983f354-77dc-4e84-969b-4c74441fa568-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.766271] env[62952]: INFO nova.compute.manager [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Terminating instance [ 937.769444] env[62952]: DEBUG nova.compute.manager [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.769444] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.770032] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7fa463-3613-4aa9-b330-1eac680b7019 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.778463] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.778752] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f893330-86df-4bab-81d7-4bfd15549489 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.783167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.786070] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Instance network_info: |[{"id": "aa4656a7-42d8-47ba-a29d-817000a4b596", "address": "fa:16:3e:14:80:1d", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa4656a7-42", "ovs_interfaceid": "aa4656a7-42d8-47ba-a29d-817000a4b596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.786070] env[62952]: DEBUG oslo_concurrency.lockutils [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] Acquired lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.786372] env[62952]: DEBUG nova.network.neutron [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Refreshing network info cache for port aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.786372] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:80:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa4656a7-42d8-47ba-a29d-817000a4b596', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.794684] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Creating folder: Project (8427597d4aae451e9375ade7c06b2bcf). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 937.797055] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7df1edaa-6b08-4568-baba-0ea35992e84f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.799086] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 937.799086] env[62952]: value = "task-1367210" [ 937.799086] env[62952]: _type = "Task" [ 937.799086] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.810667] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.812338] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Created folder: Project (8427597d4aae451e9375ade7c06b2bcf) in parent group-v290852. [ 937.812526] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Creating folder: Instances. Parent ref: group-v290904. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 937.812871] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56fa4f38-9ca4-441c-b327-ddf3d28fe759 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.823651] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Created folder: Instances in parent group-v290904. [ 937.823970] env[62952]: DEBUG oslo.service.loopingcall [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.824186] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.824393] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b057a26-92a4-4c1f-b6ba-c02aea25c679 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.849130] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.849130] env[62952]: value = "task-1367213" [ 937.849130] env[62952]: _type = "Task" [ 937.849130] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.857291] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367213, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.916910] env[62952]: INFO nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Took 30.05 seconds to build instance. [ 937.928030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.963657] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520c4a7b-1498-f25e-0a91-b5f3156a502a, 'name': SearchDatastore_Task, 'duration_secs': 0.075537} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.964157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.964575] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.964971] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d8fbc5b-e34a-4ed0-a187-72ca9d3dee56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.974517] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 937.974517] env[62952]: value = "task-1367214" [ 937.974517] env[62952]: _type = "Task" [ 937.974517] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.986294] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367214, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.123150] env[62952]: DEBUG nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.131035] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367208, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.180112] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.219276] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367209, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.313957] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367210, 'name': PowerOffVM_Task, 'duration_secs': 0.262975} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.313957] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.313957] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.313957] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2abccc2-dfd3-4853-92fb-e7d5f600829b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.362013] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367213, 'name': CreateVM_Task, 'duration_secs': 0.379362} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.362395] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.362888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.363054] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.363376] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.363634] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6713d28-5bd5-4c29-b863-1385c11e4d32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.371169] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 938.371169] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]527986e0-d7c2-d273-ce7a-aa3f8a172bd6" [ 938.371169] env[62952]: _type = "Task" [ 938.371169] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.380741] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527986e0-d7c2-d273-ce7a-aa3f8a172bd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.413023] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.413023] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.413023] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Deleting the datastore file [datastore1] 4983f354-77dc-4e84-969b-4c74441fa568 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.413023] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93069b49-fc6e-4a65-ba66-118b4cb7282e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.422572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.693s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.425260] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for the task: (returnval){ [ 938.425260] env[62952]: value = "task-1367216" [ 938.425260] env[62952]: _type = "Task" [ 938.425260] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.438355] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367216, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.476875] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ce9b45-196d-4010-9fe8-3134adcaf413 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.492958] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367214, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.496892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1477702-10f6-4a1f-a908-ca80d32d2c34 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.533795] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec90ed5-85cc-4d94-86db-9c8fae52a643 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.544381] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a311fade-5e36-45a7-b052-488a2ec50df4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.561238] env[62952]: DEBUG nova.compute.provider_tree [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.618025] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367208, 'name': ReconfigVM_Task, 'duration_secs': 0.740895} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.618132] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 93951b09-9d7c-42da-83de-0db3980d0d5d/93951b09-9d7c-42da-83de-0db3980d0d5d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.618690] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87d22a0a-5bda-4c55-bee2-0052c19ffe08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.625778] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 938.625778] env[62952]: value = "task-1367217" [ 938.625778] env[62952]: _type = "Task" [ 938.625778] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.629682] env[62952]: DEBUG nova.network.neutron [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updated VIF entry in instance network info cache for port aa4656a7-42d8-47ba-a29d-817000a4b596. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.630078] env[62952]: DEBUG nova.network.neutron [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [{"id": "aa4656a7-42d8-47ba-a29d-817000a4b596", "address": "fa:16:3e:14:80:1d", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa4656a7-42", "ovs_interfaceid": "aa4656a7-42d8-47ba-a29d-817000a4b596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.640743] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367217, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.717389] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367209, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.57635} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.718029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 94fd5287-9a8e-45e4-99e8-0bcc861f889c/94fd5287-9a8e-45e4-99e8-0bcc861f889c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.718029] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.718209] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ca76a41-f0df-4065-844b-05789ac308cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.725310] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 938.725310] env[62952]: value = "task-1367218" [ 938.725310] env[62952]: _type = "Task" [ 938.725310] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.735117] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.882846] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527986e0-d7c2-d273-ce7a-aa3f8a172bd6, 'name': SearchDatastore_Task, 'duration_secs': 0.065975} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.883167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.883399] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.883632] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.883779] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.883981] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.884262] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d801d63-1c88-4d74-b5e5-caf67121d2db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.894881] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.895133] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.895941] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-612a1b2a-d06b-44aa-986e-d0bf7fcdaee3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.900942] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 938.900942] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ea3e80-e90b-b230-1d0a-69bada326326" [ 938.900942] env[62952]: _type = "Task" [ 938.900942] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.910954] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ea3e80-e90b-b230-1d0a-69bada326326, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.929966] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 938.940501] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367216, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.986469] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367214, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678609} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.986693] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.986899] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.987156] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11226f72-dee1-4bc2-aeb3-58e951f55fa5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.993775] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 938.993775] env[62952]: value = "task-1367219" [ 938.993775] env[62952]: _type = "Task" [ 938.993775] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.001315] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367219, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.064568] env[62952]: DEBUG nova.scheduler.client.report [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.135409] env[62952]: DEBUG oslo_concurrency.lockutils [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] Releasing lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.135668] env[62952]: DEBUG nova.compute.manager [req-a72f22d9-3a23-4bb9-a49a-68e534ebcd0c req-03526511-d31b-4710-a9a3-73e47c8ac3d2 service nova] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Received event network-vif-deleted-a8d09359-772d-44c6-b657-8de81288140f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.136718] env[62952]: DEBUG nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.138584] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367217, 'name': Rename_Task, 'duration_secs': 0.422026} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.139019] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.139254] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db6974fc-42d9-4595-b8ce-d6bbbf311b69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.145862] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 939.145862] env[62952]: value = "task-1367220" [ 939.145862] env[62952]: _type = "Task" [ 939.145862] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.153282] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367220, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.160777] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.161007] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.161167] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.161352] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.161512] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.161621] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.161823] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.161980] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.162157] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.162322] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.162489] env[62952]: DEBUG nova.virt.hardware [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.163234] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6193ef03-06cf-4abb-a634-934f34dffc9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.170015] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf9b9f9-a1ae-4bae-b555-2ddada9c2753 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.183167] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.188651] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Creating folder: Project (db0e26690afb473186e344a6ecb583ec). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 939.190553] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1203ef72-bae9-4544-974c-7f3c6ca5e497 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.199924] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Created folder: Project (db0e26690afb473186e344a6ecb583ec) in parent group-v290852. [ 939.200093] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Creating folder: Instances. Parent ref: group-v290907. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 939.200314] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00ed735e-3303-4aec-bbde-3cd6fec2d52b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.210571] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Created folder: Instances in parent group-v290907. [ 939.210894] env[62952]: DEBUG oslo.service.loopingcall [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.210972] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 939.211175] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95a3a533-1dac-42d5-aafd-e7df7c1e241f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.230471] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.230471] env[62952]: value = "task-1367223" [ 939.230471] env[62952]: _type = "Task" [ 939.230471] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.236724] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087173} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.237283] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.238050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dd4404-3476-4960-a773-57bab6c6e7a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.242921] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367223, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.261917] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 94fd5287-9a8e-45e4-99e8-0bcc861f889c/94fd5287-9a8e-45e4-99e8-0bcc861f889c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.262215] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bb45519-242a-491e-af70-e61d655207d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.281968] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 939.281968] env[62952]: value = "task-1367224" [ 939.281968] env[62952]: _type = "Task" [ 939.281968] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.290466] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367224, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.411655] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ea3e80-e90b-b230-1d0a-69bada326326, 'name': SearchDatastore_Task, 'duration_secs': 0.046337} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.412449] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19adbfda-3799-4038-a4b1-6f018cb38222 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.417613] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 939.417613] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5201738b-faf7-8a6c-006d-d0143b911e20" [ 939.417613] env[62952]: _type = "Task" [ 939.417613] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.427231] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5201738b-faf7-8a6c-006d-d0143b911e20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.436521] env[62952]: DEBUG oslo_vmware.api [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Task: {'id': task-1367216, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.51958} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.436755] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.436936] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.437979] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.437979] env[62952]: INFO nova.compute.manager [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Took 1.67 seconds to destroy the instance on the hypervisor. [ 939.437979] env[62952]: DEBUG oslo.service.loopingcall [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.439505] env[62952]: DEBUG nova.compute.manager [-] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.439593] env[62952]: DEBUG nova.network.neutron [-] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 939.459446] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.504901] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367219, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071758} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.505434] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.506184] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73db2c5-1d78-4fb5-aef9-c2fb16a82417 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.531521] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.532204] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a80cbc8-09fb-4a37-8ba5-a5234eeb6e08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.551738] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 939.551738] env[62952]: value = "task-1367225" [ 939.551738] env[62952]: _type = "Task" [ 939.551738] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.561417] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367225, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.569562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.570099] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.573275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.138s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.574181] env[62952]: INFO nova.compute.claims [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.655546] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367220, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.745100] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367223, 'name': CreateVM_Task, 'duration_secs': 0.305387} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.745396] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.746050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.746319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.746688] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.747372] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-186f3703-b576-495c-8c9e-4b62ab93958c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.754448] env[62952]: DEBUG nova.compute.manager [req-8a582215-7544-4ab2-bd25-0cf8105edba0 req-8321cab5-7b4c-492a-9338-30cb5bbce36a service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Received event network-vif-deleted-1e875b4e-53fe-4d5e-b905-f34fcdcc8e86 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.754448] env[62952]: INFO nova.compute.manager [req-8a582215-7544-4ab2-bd25-0cf8105edba0 req-8321cab5-7b4c-492a-9338-30cb5bbce36a service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Neutron deleted interface 1e875b4e-53fe-4d5e-b905-f34fcdcc8e86; detaching it from the instance and deleting it from the info cache [ 939.754681] env[62952]: DEBUG nova.network.neutron [req-8a582215-7544-4ab2-bd25-0cf8105edba0 req-8321cab5-7b4c-492a-9338-30cb5bbce36a service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.758152] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 939.758152] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5263b191-694e-c5ba-77de-b93e0ffc05dd" [ 939.758152] env[62952]: _type = "Task" [ 939.758152] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.770939] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5263b191-694e-c5ba-77de-b93e0ffc05dd, 'name': SearchDatastore_Task, 'duration_secs': 0.012859} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.771963] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.772245] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.772475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.792450] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367224, 'name': ReconfigVM_Task, 'duration_secs': 0.35086} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.792759] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 94fd5287-9a8e-45e4-99e8-0bcc861f889c/94fd5287-9a8e-45e4-99e8-0bcc861f889c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.793436] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-486d54e2-a949-4376-b6ef-4c3253eab94f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.801660] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 939.801660] env[62952]: value = "task-1367226" [ 939.801660] env[62952]: _type = "Task" [ 939.801660] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.814385] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367226, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.928186] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5201738b-faf7-8a6c-006d-d0143b911e20, 'name': SearchDatastore_Task, 'duration_secs': 0.030588} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.928470] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.928788] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b/88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.929119] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.929333] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.929559] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a51cc83-7df3-4c5a-ae0a-b127ae8a7e4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.931627] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38d75254-9f22-45b9-8321-cb0ae6ac981d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.941483] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 939.941483] env[62952]: value = "task-1367227" [ 939.941483] env[62952]: _type = "Task" [ 939.941483] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.942779] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.943015] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.947017] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7251103d-2ba0-4062-9a63-827c5c89e547 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.954661] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.955938] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 939.955938] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b4e885-5250-4fcf-0798-c7d7c26344d8" [ 939.955938] env[62952]: _type = "Task" [ 939.955938] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.963787] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b4e885-5250-4fcf-0798-c7d7c26344d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.061307] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367225, 'name': ReconfigVM_Task, 'duration_secs': 0.345169} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.061743] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e/34f5995f-692c-4f0d-8b15-0d388df7e34e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.062504] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c2b4679-6ec4-4f2d-953c-06bb181f1b00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.069585] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 940.069585] env[62952]: value = "task-1367228" [ 940.069585] env[62952]: _type = "Task" [ 940.069585] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.078263] env[62952]: DEBUG nova.compute.utils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.081442] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367228, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.081902] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.082102] env[62952]: DEBUG nova.network.neutron [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 940.145071] env[62952]: DEBUG nova.policy [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f24a30e1160a43d5be6cde39b8f79edd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '094ca56818604bee9c1f75d1493ab3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.157028] env[62952]: DEBUG oslo_vmware.api [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367220, 'name': PowerOnVM_Task, 'duration_secs': 0.724742} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.157260] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.157476] env[62952]: INFO nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Took 10.20 seconds to spawn the instance on the hypervisor. [ 940.158390] env[62952]: DEBUG nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.159235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297ee7af-875d-4b34-b72c-8c6692e17ddc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.212519] env[62952]: DEBUG nova.network.neutron [-] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.257343] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d666d6ed-08c9-493d-a982-2ced5603b037 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.270521] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c88e58-d138-45de-a49e-e3a7d6aec519 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.305379] env[62952]: DEBUG nova.compute.manager [req-8a582215-7544-4ab2-bd25-0cf8105edba0 req-8321cab5-7b4c-492a-9338-30cb5bbce36a service nova] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Detach interface failed, port_id=1e875b4e-53fe-4d5e-b905-f34fcdcc8e86, reason: Instance 4983f354-77dc-4e84-969b-4c74441fa568 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 940.316448] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367226, 'name': Rename_Task, 'duration_secs': 0.149978} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.316727] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.316975] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be82a740-89f0-4e92-8bb2-de0deb228ee0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.327936] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 940.327936] env[62952]: value = "task-1367229" [ 940.327936] env[62952]: _type = "Task" [ 940.327936] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.340927] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.443831] env[62952]: DEBUG nova.network.neutron [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Successfully created port: 034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.458300] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367227, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.468158] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b4e885-5250-4fcf-0798-c7d7c26344d8, 'name': SearchDatastore_Task, 'duration_secs': 0.011949} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.469074] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e853c324-84ca-4b81-a21a-8d7a1235a99d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.477528] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 940.477528] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52998e2a-91a6-f88e-cd82-f7e7a31f21f1" [ 940.477528] env[62952]: _type = "Task" [ 940.477528] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.490471] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52998e2a-91a6-f88e-cd82-f7e7a31f21f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.580012] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367228, 'name': Rename_Task, 'duration_secs': 0.139105} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.580368] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.580647] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3ccf094-9213-4a3f-92f5-1493c9a08a3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.582570] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.598259] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 940.598259] env[62952]: value = "task-1367230" [ 940.598259] env[62952]: _type = "Task" [ 940.598259] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.607281] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367230, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.683290] env[62952]: INFO nova.compute.manager [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Took 31.88 seconds to build instance. [ 940.717250] env[62952]: INFO nova.compute.manager [-] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Took 1.28 seconds to deallocate network for instance. [ 940.838710] env[62952]: DEBUG oslo_vmware.api [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367229, 'name': PowerOnVM_Task, 'duration_secs': 0.497652} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.841352] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.841596] env[62952]: INFO nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Took 8.50 seconds to spawn the instance on the hypervisor. [ 940.841766] env[62952]: DEBUG nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.842992] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6804160b-91db-4298-b9fd-6a550e9556c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.885508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3736d7ad-17a5-4eb5-9c3c-6443862e9548 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.893967] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4a21f9-edc4-4fcb-b057-20834b0c0554 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.925146] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39e7ee0-7ea6-498f-b2af-2b863334db0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.934287] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d2027d-2f7d-4bed-80a7-ec54077a82de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.950624] env[62952]: DEBUG nova.compute.provider_tree [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.960755] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367227, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.624618} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.961016] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b/88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.961422] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.961725] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d96d45a-a5ca-4b4e-be00-9b7647e94045 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.971804] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 940.971804] env[62952]: value = "task-1367231" [ 940.971804] env[62952]: _type = "Task" [ 940.971804] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.980444] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.990020] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52998e2a-91a6-f88e-cd82-f7e7a31f21f1, 'name': SearchDatastore_Task, 'duration_secs': 0.056244} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.990296] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.990550] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.990805] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f1fdf71-f1b0-450a-a44d-a568064a8de0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.002058] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 941.002058] env[62952]: value = "task-1367232" [ 941.002058] env[62952]: _type = "Task" [ 941.002058] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.012686] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367232, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.109097] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367230, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.185666] env[62952]: DEBUG oslo_concurrency.lockutils [None req-25a7d206-8a4f-42c7-a4ce-01e954328a59 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.422s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.227813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.365075] env[62952]: INFO nova.compute.manager [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Took 30.64 seconds to build instance. [ 941.455898] env[62952]: DEBUG nova.scheduler.client.report [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.484144] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.22482} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.484470] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.485331] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7848b7db-93c8-48e4-b8b6-f8677d35c7c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.508071] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b/88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.508475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "7621d6a6-27aa-45d3-80fa-65e957519a1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.508695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.508884] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "7621d6a6-27aa-45d3-80fa-65e957519a1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.509121] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.509316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.510845] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54b3f9da-fc8c-4eb5-91e2-758479b7c6f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.527663] env[62952]: INFO nova.compute.manager [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Terminating instance [ 941.532922] env[62952]: DEBUG nova.compute.manager [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.533175] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.534447] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdc54c3-c681-468f-9a94-ccce402c1a30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.540515] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 941.540515] env[62952]: value = "task-1367233" [ 941.540515] env[62952]: _type = "Task" [ 941.540515] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.544114] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367232, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.549347] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.549622] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fff2c66-1fed-4f81-b46b-026ec6e900ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.557489] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367233, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.558796] env[62952]: DEBUG oslo_vmware.api [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 941.558796] env[62952]: value = "task-1367234" [ 941.558796] env[62952]: _type = "Task" [ 941.558796] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.568065] env[62952]: DEBUG oslo_vmware.api [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.598566] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.610828] env[62952]: DEBUG oslo_vmware.api [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367230, 'name': PowerOnVM_Task, 'duration_secs': 0.979249} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.611108] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.611319] env[62952]: DEBUG nova.compute.manager [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.612206] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d03a4d9-31b7-4046-8e74-26059647726c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.627272] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.627520] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.627675] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.627852] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.627997] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.628176] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.628476] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.628748] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.628941] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.629127] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.629314] env[62952]: DEBUG nova.virt.hardware [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.630546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c56ff56-f808-421a-ba31-efd19dc80f6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.639113] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1af4aa5-6502-4685-8e65-f51453d93070 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.657404] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "93951b09-9d7c-42da-83de-0db3980d0d5d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.657637] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.657830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "93951b09-9d7c-42da-83de-0db3980d0d5d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.658017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.658198] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.660204] env[62952]: INFO nova.compute.manager [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Terminating instance [ 941.661991] env[62952]: DEBUG nova.compute.manager [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.662195] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.662957] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105c5773-f142-4b15-9551-326d4d7d66e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.672165] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.672478] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5f801d7-595b-4e7c-8404-f105b4b4dbb0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.681872] env[62952]: DEBUG oslo_vmware.api [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 941.681872] env[62952]: value = "task-1367235" [ 941.681872] env[62952]: _type = "Task" [ 941.681872] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.699614] env[62952]: DEBUG oslo_vmware.api [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367235, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.867271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d3fb43b7-7449-42e2-8db1-22ea289b7a22 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.830s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.961235] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.961839] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 941.967215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.501s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.970433] env[62952]: INFO nova.compute.claims [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.974830] env[62952]: DEBUG nova.compute.manager [req-085ea8d6-886d-4188-9625-afb92dd7c8b9 req-0c974c36-7d90-4f1b-9f07-35baa7671a27 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Received event network-vif-plugged-034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.975083] env[62952]: DEBUG oslo_concurrency.lockutils [req-085ea8d6-886d-4188-9625-afb92dd7c8b9 req-0c974c36-7d90-4f1b-9f07-35baa7671a27 service nova] Acquiring lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.975281] env[62952]: DEBUG oslo_concurrency.lockutils [req-085ea8d6-886d-4188-9625-afb92dd7c8b9 req-0c974c36-7d90-4f1b-9f07-35baa7671a27 service nova] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.975444] env[62952]: DEBUG oslo_concurrency.lockutils [req-085ea8d6-886d-4188-9625-afb92dd7c8b9 req-0c974c36-7d90-4f1b-9f07-35baa7671a27 service nova] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.975608] env[62952]: DEBUG nova.compute.manager [req-085ea8d6-886d-4188-9625-afb92dd7c8b9 req-0c974c36-7d90-4f1b-9f07-35baa7671a27 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] No waiting events found dispatching network-vif-plugged-034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.975782] env[62952]: WARNING nova.compute.manager [req-085ea8d6-886d-4188-9625-afb92dd7c8b9 req-0c974c36-7d90-4f1b-9f07-35baa7671a27 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Received unexpected event network-vif-plugged-034e9fd2-98bf-4d19-a01e-627fcfc75b89 for instance with vm_state building and task_state spawning. [ 942.024825] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367232, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.059274] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.068748] env[62952]: DEBUG oslo_vmware.api [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367234, 'name': PowerOffVM_Task, 'duration_secs': 0.348579} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.069181] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.069321] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.069559] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a01a638b-cb63-43ec-a6e0-249d076a112e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.087282] env[62952]: DEBUG nova.network.neutron [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Successfully updated port: 034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.132200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.168517] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.168757] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.168942] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleting the datastore file [datastore1] 7621d6a6-27aa-45d3-80fa-65e957519a1f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.169226] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66c6b158-6c68-462a-8061-590318d796ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.177137] env[62952]: DEBUG oslo_vmware.api [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 942.177137] env[62952]: value = "task-1367237" [ 942.177137] env[62952]: _type = "Task" [ 942.177137] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.186392] env[62952]: DEBUG oslo_vmware.api [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367237, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.195793] env[62952]: DEBUG oslo_vmware.api [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367235, 'name': PowerOffVM_Task, 'duration_secs': 0.282818} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.196111] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.196289] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.196762] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72ed132c-2890-4f36-809c-37fae0fbfb6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.278318] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.278516] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.278819] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleting the datastore file [datastore1] 93951b09-9d7c-42da-83de-0db3980d0d5d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.279161] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7ffc2f9-d8d4-4162-b243-d0d9a612e71d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.286745] env[62952]: DEBUG oslo_vmware.api [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 942.286745] env[62952]: value = "task-1367239" [ 942.286745] env[62952]: _type = "Task" [ 942.286745] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.298106] env[62952]: DEBUG oslo_vmware.api [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.477373] env[62952]: DEBUG nova.compute.utils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.480657] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.480807] env[62952]: DEBUG nova.network.neutron [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 942.527023] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367232, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.037947} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.527023] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.527023] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.527023] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2bf32eb-5027-496a-abae-3dd13ef7cd36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.536835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "34f5995f-692c-4f0d-8b15-0d388df7e34e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.536835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.536835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "34f5995f-692c-4f0d-8b15-0d388df7e34e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.536835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.537160] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.542512] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 942.542512] env[62952]: value = "task-1367240" [ 942.542512] env[62952]: _type = "Task" [ 942.542512] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.543000] env[62952]: INFO nova.compute.manager [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Terminating instance [ 942.549134] env[62952]: DEBUG nova.compute.manager [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.549420] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.550822] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344e4e2f-16e0-46e7-93b6-f70270ec4f91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.564190] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367240, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.566015] env[62952]: DEBUG nova.policy [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a952cf355ac14bf7b75c5d185c440982', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88312f85f00741aba154b9744055c17d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.572682] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.572923] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367233, 'name': ReconfigVM_Task, 'duration_secs': 0.879926} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.573533] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8811cf2-e0d9-46a8-aa99-b2d74bf524eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.575842] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b/88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.576932] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6eb46dc5-46b5-410e-bc8e-f8b70954a5d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.587539] env[62952]: DEBUG oslo_vmware.api [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 942.587539] env[62952]: value = "task-1367241" [ 942.587539] env[62952]: _type = "Task" [ 942.587539] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.589048] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 942.589048] env[62952]: value = "task-1367242" [ 942.589048] env[62952]: _type = "Task" [ 942.589048] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.593327] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.593534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.593745] env[62952]: DEBUG nova.network.neutron [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.606547] env[62952]: DEBUG oslo_vmware.api [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367241, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.612050] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367242, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.689548] env[62952]: DEBUG oslo_vmware.api [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367237, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21194} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.689926] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.690168] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.691020] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.691020] env[62952]: INFO nova.compute.manager [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 942.691020] env[62952]: DEBUG oslo.service.loopingcall [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.691292] env[62952]: DEBUG nova.compute.manager [-] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.691292] env[62952]: DEBUG nova.network.neutron [-] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.804071] env[62952]: DEBUG oslo_vmware.api [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216623} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.804275] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.804614] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.804719] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.805714] env[62952]: INFO nova.compute.manager [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 942.805714] env[62952]: DEBUG oslo.service.loopingcall [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.805714] env[62952]: DEBUG nova.compute.manager [-] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.805714] env[62952]: DEBUG nova.network.neutron [-] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.986032] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281cb531-4502-4c20-a1ed-f0435f3fa3cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.990496] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.013205] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Suspending the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 943.016609] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-316be039-cabd-46e8-afc3-788ebaa042c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.032761] env[62952]: DEBUG oslo_vmware.api [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 943.032761] env[62952]: value = "task-1367243" [ 943.032761] env[62952]: _type = "Task" [ 943.032761] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.045821] env[62952]: DEBUG oslo_vmware.api [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367243, 'name': SuspendVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.059887] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310628} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.060436] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.061368] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c142c6-2137-4398-a185-0a95c3dfe636 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.082313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "3bd9c356-42d6-450f-9069-b066952d9524" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.082313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "3bd9c356-42d6-450f-9069-b066952d9524" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.094664] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.096272] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d16063b-025e-4782-8484-b14c1c956686 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.124394] env[62952]: DEBUG nova.network.neutron [-] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.134604] env[62952]: DEBUG oslo_vmware.api [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367241, 'name': PowerOffVM_Task, 'duration_secs': 0.312945} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.140361] env[62952]: DEBUG nova.compute.manager [req-f4797a54-c319-482d-92f8-f105c3c8c1c4 req-adff8f6a-1f27-4224-8eaa-e9a8515ab2c4 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Received event network-vif-deleted-fd33b4df-02a6-48ec-8118-6ffc34a8e633 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.141025] env[62952]: INFO nova.compute.manager [req-f4797a54-c319-482d-92f8-f105c3c8c1c4 req-adff8f6a-1f27-4224-8eaa-e9a8515ab2c4 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Neutron deleted interface fd33b4df-02a6-48ec-8118-6ffc34a8e633; detaching it from the instance and deleting it from the info cache [ 943.141025] env[62952]: DEBUG nova.network.neutron [req-f4797a54-c319-482d-92f8-f105c3c8c1c4 req-adff8f6a-1f27-4224-8eaa-e9a8515ab2c4 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.141913] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.141913] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.143024] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 943.143024] env[62952]: value = "task-1367244" [ 943.143024] env[62952]: _type = "Task" [ 943.143024] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.143024] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367242, 'name': Rename_Task, 'duration_secs': 0.361473} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.145933] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97ac9941-55b4-4df7-92d2-376909f15e90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.147715] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.151962] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7ab5eb6-b615-40f1-b57f-06cc34364064 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.162815] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 943.162815] env[62952]: value = "task-1367245" [ 943.162815] env[62952]: _type = "Task" [ 943.162815] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.170596] env[62952]: DEBUG nova.network.neutron [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.179707] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.223038] env[62952]: DEBUG nova.network.neutron [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Successfully created port: 104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.237464] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 943.237464] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 943.237464] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore2] 34f5995f-692c-4f0d-8b15-0d388df7e34e {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.237464] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9015e20d-d630-445f-ba21-e17278e4b94f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.249870] env[62952]: DEBUG oslo_vmware.api [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 943.249870] env[62952]: value = "task-1367247" [ 943.249870] env[62952]: _type = "Task" [ 943.249870] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.259531] env[62952]: DEBUG oslo_vmware.api [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.355118] env[62952]: DEBUG nova.network.neutron [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updating instance_info_cache with network_info: [{"id": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "address": "fa:16:3e:f0:ce:a3", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034e9fd2-98", "ovs_interfaceid": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.385899] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7382b0e8-1e1f-4ef8-b158-062fd4a37126 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.393992] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf1b476-00ca-489c-8227-f44f9a28a5b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.426402] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65c4caf-fc70-49e8-9525-46e6f13e4cce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.434943] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ffb800-5e56-4569-8b28-f14b53cb8208 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.450218] env[62952]: DEBUG nova.compute.provider_tree [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.543927] env[62952]: DEBUG oslo_vmware.api [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367243, 'name': SuspendVM_Task} progress is 41%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.596725] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 943.627919] env[62952]: INFO nova.compute.manager [-] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Took 0.82 seconds to deallocate network for instance. [ 943.644666] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-315d4466-3a1b-4dac-b73c-63d7d7ed4305 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.657621] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5281d6f-2efb-43d0-96e9-50d40afeb914 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.675761] env[62952]: DEBUG nova.network.neutron [-] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.676629] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367244, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.687926] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367245, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.697529] env[62952]: DEBUG nova.compute.manager [req-f4797a54-c319-482d-92f8-f105c3c8c1c4 req-adff8f6a-1f27-4224-8eaa-e9a8515ab2c4 service nova] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Detach interface failed, port_id=fd33b4df-02a6-48ec-8118-6ffc34a8e633, reason: Instance 93951b09-9d7c-42da-83de-0db3980d0d5d could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 943.763138] env[62952]: DEBUG oslo_vmware.api [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316894} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.763138] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.763265] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.763424] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.763637] env[62952]: INFO nova.compute.manager [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Took 1.21 seconds to destroy the instance on the hypervisor. [ 943.763972] env[62952]: DEBUG oslo.service.loopingcall [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.764830] env[62952]: DEBUG nova.compute.manager [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 943.764830] env[62952]: DEBUG nova.network.neutron [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 943.857764] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.858949] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Instance network_info: |[{"id": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "address": "fa:16:3e:f0:ce:a3", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034e9fd2-98", "ovs_interfaceid": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.859656] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:ce:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97113f46-d648-4613-b233-069acba18198', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '034e9fd2-98bf-4d19-a01e-627fcfc75b89', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.871318] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating folder: Project (094ca56818604bee9c1f75d1493ab3af). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 943.871318] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58f7e2ae-74ac-40e4-95df-a02f0cf0a09f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.889999] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Created folder: Project (094ca56818604bee9c1f75d1493ab3af) in parent group-v290852. [ 943.890448] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating folder: Instances. Parent ref: group-v290910. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 943.890778] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd7622ce-769f-4c82-b379-8e9958ac2ac0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.909030] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Created folder: Instances in parent group-v290910. [ 943.909030] env[62952]: DEBUG oslo.service.loopingcall [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.909030] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.909030] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11b99282-c0ea-4802-8b0f-88805d4a273e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.941027] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.941027] env[62952]: value = "task-1367250" [ 943.941027] env[62952]: _type = "Task" [ 943.941027] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.949460] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367250, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.956176] env[62952]: DEBUG nova.scheduler.client.report [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.015020] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.048038] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.048038] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.048038] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.048310] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.048310] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.048550] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.048901] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.049211] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.049515] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.049801] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.050108] env[62952]: DEBUG nova.virt.hardware [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.051552] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937b0581-f743-4911-b504-1db54d3055c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.059859] env[62952]: DEBUG oslo_vmware.api [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367243, 'name': SuspendVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.066150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1cde38-e238-428a-8d0c-e6dd882b4c8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.088796] env[62952]: DEBUG nova.compute.manager [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Received event network-changed-034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.089036] env[62952]: DEBUG nova.compute.manager [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Refreshing instance network info cache due to event network-changed-034e9fd2-98bf-4d19-a01e-627fcfc75b89. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.089253] env[62952]: DEBUG oslo_concurrency.lockutils [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] Acquiring lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.089418] env[62952]: DEBUG oslo_concurrency.lockutils [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] Acquired lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.089592] env[62952]: DEBUG nova.network.neutron [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Refreshing network info cache for port 034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.127491] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.135240] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.158421] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367244, 'name': ReconfigVM_Task, 'duration_secs': 0.731613} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.158706] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Reconfigured VM instance instance-00000047 to attach disk [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.159373] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56b0e892-e3db-4b41-9268-d42c881b4d8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.168602] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 944.168602] env[62952]: value = "task-1367251" [ 944.168602] env[62952]: _type = "Task" [ 944.168602] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.179559] env[62952]: INFO nova.compute.manager [-] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Took 1.49 seconds to deallocate network for instance. [ 944.179841] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367251, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.194033] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367245, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.449821] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367250, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.458825] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.459415] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.462872] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.614s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.463093] env[62952]: DEBUG nova.objects.instance [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lazy-loading 'resources' on Instance uuid 5914a35a-0934-4f8f-81e8-d91bc690a9cb {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.545180] env[62952]: DEBUG oslo_vmware.api [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367243, 'name': SuspendVM_Task, 'duration_secs': 1.095195} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.545556] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Suspended the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 944.545701] env[62952]: DEBUG nova.compute.manager [None req-f3d271c3-7885-49fa-a94f-525429c04f60 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.546587] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c65a628-ac46-4ada-ad53-260ca2aa651a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.664600] env[62952]: DEBUG nova.network.neutron [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.678725] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367251, 'name': Rename_Task, 'duration_secs': 0.249271} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.679180] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.679626] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21d6bdf9-4937-4812-8cfd-e3e065a9fde6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.691426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.692087] env[62952]: DEBUG oslo_vmware.api [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367245, 'name': PowerOnVM_Task, 'duration_secs': 1.52881} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.693446] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.693779] env[62952]: INFO nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Took 9.92 seconds to spawn the instance on the hypervisor. [ 944.694085] env[62952]: DEBUG nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.694742] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 944.694742] env[62952]: value = "task-1367252" [ 944.694742] env[62952]: _type = "Task" [ 944.694742] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.695918] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956aa06e-9e56-48e3-9b54-aa6b75031d00 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.711986] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367252, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.783955] env[62952]: DEBUG nova.network.neutron [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Successfully updated port: 104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.889395] env[62952]: DEBUG nova.network.neutron [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updated VIF entry in instance network info cache for port 034e9fd2-98bf-4d19-a01e-627fcfc75b89. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.889801] env[62952]: DEBUG nova.network.neutron [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updating instance_info_cache with network_info: [{"id": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "address": "fa:16:3e:f0:ce:a3", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034e9fd2-98", "ovs_interfaceid": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.952989] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367250, 'name': CreateVM_Task, 'duration_secs': 0.546474} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.952989] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 944.953208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.953369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.953696] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.954075] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b1df017-e2f6-4285-8829-c6cbcceef82a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.958678] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 944.958678] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523cfd2c-cc86-f021-71f3-d067ec2834f2" [ 944.958678] env[62952]: _type = "Task" [ 944.958678] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.967957] env[62952]: DEBUG nova.compute.utils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.971712] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523cfd2c-cc86-f021-71f3-d067ec2834f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.972397] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 944.972568] env[62952]: DEBUG nova.network.neutron [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 945.022164] env[62952]: DEBUG nova.policy [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8171575c58594e7ebc37af7c4c858c10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c451895447b473eb18675f6bc843081', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.170575] env[62952]: INFO nova.compute.manager [-] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Took 1.41 seconds to deallocate network for instance. [ 945.180821] env[62952]: DEBUG nova.compute.manager [req-5bb08e2e-4c3b-4d46-952c-13a18a09102c req-837bdac7-66ab-4ce2-b75b-8f1a5f238ef3 service nova] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Received event network-vif-deleted-feeba8e0-a5e3-4e76-9991-96695e8032eb {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.216203] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.228353] env[62952]: INFO nova.compute.manager [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Took 33.95 seconds to build instance. [ 945.265106] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44b5cbd-0add-4363-be0c-41ed5eff4fdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.272646] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26012da-d696-4334-a6a5-fbe822850c5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.305469] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "refresh_cache-a88447f6-6a5a-490d-bf44-11c9237fa5cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.305614] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquired lock "refresh_cache-a88447f6-6a5a-490d-bf44-11c9237fa5cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.305771] env[62952]: DEBUG nova.network.neutron [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.307921] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43dfbb7-76a3-458c-9d35-e638f09fa2ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.316354] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c0bb22-cfb6-47a8-a497-ecf449aad7ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.330985] env[62952]: DEBUG nova.compute.provider_tree [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.392593] env[62952]: DEBUG oslo_concurrency.lockutils [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] Releasing lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.392852] env[62952]: DEBUG nova.compute.manager [req-39914846-a9ad-46b9-8f1b-10afbb125571 req-17317822-849e-491e-bb5e-a05978e6b894 service nova] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Received event network-vif-deleted-c1a7ed7e-0067-4158-95c8-bdc2c8104c1d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.402303] env[62952]: DEBUG nova.network.neutron [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Successfully created port: 4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.472017] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523cfd2c-cc86-f021-71f3-d067ec2834f2, 'name': SearchDatastore_Task, 'duration_secs': 0.019583} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.472823] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.475902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.476144] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.476374] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.476518] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.482394] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.482394] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d9f5a20-b8eb-48c1-831b-4adac311c38f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.492390] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.492560] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.493300] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-384fb2cb-f1d3-4966-ae7a-e35b1b377ff6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.499113] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 945.499113] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c25dae-3590-33ee-64ce-b643f2f6e67b" [ 945.499113] env[62952]: _type = "Task" [ 945.499113] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.506985] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c25dae-3590-33ee-64ce-b643f2f6e67b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.685488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.712458] env[62952]: DEBUG oslo_vmware.api [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367252, 'name': PowerOnVM_Task, 'duration_secs': 1.02069} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.712458] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.712458] env[62952]: INFO nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Took 6.58 seconds to spawn the instance on the hypervisor. [ 945.712458] env[62952]: DEBUG nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.713574] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f85ba04-60c2-4950-811c-8ce57022cbc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.730461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9a2f5826-52d2-48ba-a1a6-1cedb8dae09c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.742s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.834573] env[62952]: DEBUG nova.scheduler.client.report [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.854538] env[62952]: DEBUG nova.network.neutron [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.011541] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c25dae-3590-33ee-64ce-b643f2f6e67b, 'name': SearchDatastore_Task, 'duration_secs': 0.009185} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.012442] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-925f5e51-3ee4-480c-95e8-031178f7bdca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.020019] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 946.020019] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52837b10-54d0-0908-f233-26b6942db485" [ 946.020019] env[62952]: _type = "Task" [ 946.020019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.029847] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52837b10-54d0-0908-f233-26b6942db485, 'name': SearchDatastore_Task, 'duration_secs': 0.00818} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.030115] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.030640] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 5ace8d6f-49ab-4486-bee1-04cd24f7f430/5ace8d6f-49ab-4486-bee1-04cd24f7f430.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.030894] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29a33b21-4690-46c9-8bb5-113e020f707f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.039416] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 946.039416] env[62952]: value = "task-1367253" [ 946.039416] env[62952]: _type = "Task" [ 946.039416] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.051756] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.051756] env[62952]: DEBUG nova.network.neutron [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Updating instance_info_cache with network_info: [{"id": "104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2", "address": "fa:16:3e:78:ff:8a", "network": {"id": "a70cd245-0c2e-4124-bfa1-3802b7fd1de8", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-2133534156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88312f85f00741aba154b9744055c17d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap104e8b8a-aa", "ovs_interfaceid": "104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.122918] env[62952]: DEBUG nova.compute.manager [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Received event network-vif-plugged-104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.123199] env[62952]: DEBUG oslo_concurrency.lockutils [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] Acquiring lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.123440] env[62952]: DEBUG oslo_concurrency.lockutils [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.123673] env[62952]: DEBUG oslo_concurrency.lockutils [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.123904] env[62952]: DEBUG nova.compute.manager [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] No waiting events found dispatching network-vif-plugged-104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.124262] env[62952]: WARNING nova.compute.manager [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Received unexpected event network-vif-plugged-104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 for instance with vm_state building and task_state spawning. [ 946.124377] env[62952]: DEBUG nova.compute.manager [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Received event network-changed-104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.124526] env[62952]: DEBUG nova.compute.manager [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Refreshing instance network info cache due to event network-changed-104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.124730] env[62952]: DEBUG oslo_concurrency.lockutils [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] Acquiring lock "refresh_cache-a88447f6-6a5a-490d-bf44-11c9237fa5cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.231126] env[62952]: INFO nova.compute.manager [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Took 26.68 seconds to build instance. [ 946.343736] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.345316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.370s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.347872] env[62952]: INFO nova.compute.claims [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.383617] env[62952]: INFO nova.scheduler.client.report [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Deleted allocations for instance 5914a35a-0934-4f8f-81e8-d91bc690a9cb [ 946.485501] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.516021] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.516021] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.516021] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.516287] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.516287] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.516287] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.516287] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.516287] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.516428] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.516428] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.516428] env[62952]: DEBUG nova.virt.hardware [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.517150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd3bf9b-c140-473b-b157-c7c599bde08a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.526472] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc18659d-9548-4104-b8e1-0da14c066cd6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.547842] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454942} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.548137] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 5ace8d6f-49ab-4486-bee1-04cd24f7f430/5ace8d6f-49ab-4486-bee1-04cd24f7f430.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 946.548355] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 946.548598] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-177b79da-d77b-44ba-87bf-39b566a440e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.554009] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Releasing lock "refresh_cache-a88447f6-6a5a-490d-bf44-11c9237fa5cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.554345] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Instance network_info: |[{"id": "104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2", "address": "fa:16:3e:78:ff:8a", "network": {"id": "a70cd245-0c2e-4124-bfa1-3802b7fd1de8", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-2133534156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88312f85f00741aba154b9744055c17d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap104e8b8a-aa", "ovs_interfaceid": "104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.556851] env[62952]: DEBUG oslo_concurrency.lockutils [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] Acquired lock "refresh_cache-a88447f6-6a5a-490d-bf44-11c9237fa5cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.557042] env[62952]: DEBUG nova.network.neutron [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Refreshing network info cache for port 104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.558309] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:ff:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '99be9a5e-b3f9-4e6c-83d5-df11f817847d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.570734] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Creating folder: Project (88312f85f00741aba154b9744055c17d). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.570999] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 946.570999] env[62952]: value = "task-1367254" [ 946.570999] env[62952]: _type = "Task" [ 946.570999] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.573914] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2cf6e21f-fcf0-4b43-aa22-d53da9e8f0df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.585146] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367254, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.587438] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Created folder: Project (88312f85f00741aba154b9744055c17d) in parent group-v290852. [ 946.587657] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Creating folder: Instances. Parent ref: group-v290913. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.587894] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4622dd95-8e7d-478f-b591-93b096c007eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.599589] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Created folder: Instances in parent group-v290913. [ 946.599860] env[62952]: DEBUG oslo.service.loopingcall [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.600104] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.600498] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb81bbcc-3bd5-4c1a-a714-aa7f18f8b34e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.626977] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.626977] env[62952]: value = "task-1367257" [ 946.626977] env[62952]: _type = "Task" [ 946.626977] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.637343] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367257, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.735444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-afe1da2b-addc-4993-8e41-f9d7e30c92ad tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "a3c01d4d-9c66-4441-95e8-87998782cc02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.058s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.809218] env[62952]: DEBUG nova.network.neutron [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Updated VIF entry in instance network info cache for port 104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.809593] env[62952]: DEBUG nova.network.neutron [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Updating instance_info_cache with network_info: [{"id": "104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2", "address": "fa:16:3e:78:ff:8a", "network": {"id": "a70cd245-0c2e-4124-bfa1-3802b7fd1de8", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-2133534156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88312f85f00741aba154b9744055c17d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "99be9a5e-b3f9-4e6c-83d5-df11f817847d", "external-id": "nsx-vlan-transportzone-566", "segmentation_id": 566, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap104e8b8a-aa", "ovs_interfaceid": "104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.866127] env[62952]: DEBUG nova.compute.manager [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.867671] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d0a519-7d75-4dd3-bda5-a40b1385a4e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.891857] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3fcb5547-9218-4509-9a8e-4455859992ae tempest-ServersAaction247Test-1095029802 tempest-ServersAaction247Test-1095029802-project-member] Lock "5914a35a-0934-4f8f-81e8-d91bc690a9cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.923s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.932946] env[62952]: DEBUG nova.network.neutron [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Successfully updated port: 4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.087593] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367254, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081081} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.087844] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.088660] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1827a1-183d-4599-9a4f-58281d820a9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.110845] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 5ace8d6f-49ab-4486-bee1-04cd24f7f430/5ace8d6f-49ab-4486-bee1-04cd24f7f430.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.111174] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52dd6444-c57b-44d6-8424-8f61415537d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.125363] env[62952]: INFO nova.compute.manager [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Rebuilding instance [ 947.133739] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 947.133739] env[62952]: value = "task-1367258" [ 947.133739] env[62952]: _type = "Task" [ 947.133739] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.150801] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367258, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.151069] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367257, 'name': CreateVM_Task, 'duration_secs': 0.409494} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.155176] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.156271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.156439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.156792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.157088] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e6a3bbc-b83b-4f24-9032-dbe0408450c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.162816] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 947.162816] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523d3943-3a11-6c4e-5c7f-cc4f38a19036" [ 947.162816] env[62952]: _type = "Task" [ 947.162816] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.171145] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523d3943-3a11-6c4e-5c7f-cc4f38a19036, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.179262] env[62952]: DEBUG nova.compute.manager [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.179936] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34db2e0-2e19-4eca-af55-c34c7e17e2e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.215680] env[62952]: DEBUG nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Received event network-changed-aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.215870] env[62952]: DEBUG nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Refreshing instance network info cache due to event network-changed-aa4656a7-42d8-47ba-a29d-817000a4b596. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.216098] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Acquiring lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.216239] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Acquired lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.216405] env[62952]: DEBUG nova.network.neutron [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Refreshing network info cache for port aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.312980] env[62952]: DEBUG oslo_concurrency.lockutils [req-8aeca1fd-794e-45ca-b7ac-1bf511a21e70 req-c839d632-f4bd-45bf-a18f-8e061fe649a7 service nova] Releasing lock "refresh_cache-a88447f6-6a5a-490d-bf44-11c9237fa5cb" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.379266] env[62952]: INFO nova.compute.manager [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] instance snapshotting [ 947.379478] env[62952]: WARNING nova.compute.manager [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 947.382193] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dc6b4b-2cbd-45fe-b09f-cbb9b0a152ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.402010] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0dfea2-12ef-439f-8273-90700b3da3b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.436042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.436230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.436358] env[62952]: DEBUG nova.network.neutron [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.622819] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb23190-e70a-42a4-93d1-d735e97e98e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.633058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd56aaa7-fd95-4c0a-915b-b2610ff749b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.672998] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c9adbc-f665-4fac-97ae-460afa3fd008 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.679887] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.686525] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523d3943-3a11-6c4e-5c7f-cc4f38a19036, 'name': SearchDatastore_Task, 'duration_secs': 0.009279} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.688699] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.689316] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.689316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.689513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.689624] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.690859] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.691408] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0691e994-f9c6-4166-8fb7-7176c59ebd7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.693884] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c75c2c-674e-45f6-bfe2-1b59298f0572 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.697839] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afe48780-2ee9-414a-8fe1-f373f7707471 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.710272] env[62952]: DEBUG nova.compute.provider_tree [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.716435] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 947.716435] env[62952]: value = "task-1367259" [ 947.716435] env[62952]: _type = "Task" [ 947.716435] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.716768] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.716910] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.718784] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b9361c2-e30e-43f7-b33f-5f9d38825398 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.734555] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367259, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.734791] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 947.734791] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5222bdce-f8ca-74e3-8561-762594f502b7" [ 947.734791] env[62952]: _type = "Task" [ 947.734791] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.744462] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5222bdce-f8ca-74e3-8561-762594f502b7, 'name': SearchDatastore_Task, 'duration_secs': 0.009851} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.745321] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01f7d198-9e66-4d60-b3dc-4567e93beb1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.752500] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 947.752500] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a796e7-6ff0-b092-bc49-172608abec54" [ 947.752500] env[62952]: _type = "Task" [ 947.752500] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.765726] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a796e7-6ff0-b092-bc49-172608abec54, 'name': SearchDatastore_Task} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.766139] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.766500] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] a88447f6-6a5a-490d-bf44-11c9237fa5cb/a88447f6-6a5a-490d-bf44-11c9237fa5cb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.766758] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba065f20-c710-47b8-ac53-ea3add4cb732 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.775847] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 947.775847] env[62952]: value = "task-1367260" [ 947.775847] env[62952]: _type = "Task" [ 947.775847] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.792115] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.914219] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 947.914614] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d20016bf-9ce8-4c0c-ae71-ce1fa2df185f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.923973] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 947.923973] env[62952]: value = "task-1367261" [ 947.923973] env[62952]: _type = "Task" [ 947.923973] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.932728] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367261, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.972182] env[62952]: DEBUG nova.network.neutron [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.054871] env[62952]: DEBUG nova.network.neutron [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updated VIF entry in instance network info cache for port aa4656a7-42d8-47ba-a29d-817000a4b596. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.056017] env[62952]: DEBUG nova.network.neutron [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [{"id": "aa4656a7-42d8-47ba-a29d-817000a4b596", "address": "fa:16:3e:14:80:1d", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa4656a7-42", "ovs_interfaceid": "aa4656a7-42d8-47ba-a29d-817000a4b596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.148973] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367258, 'name': ReconfigVM_Task, 'duration_secs': 0.837048} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.149372] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 5ace8d6f-49ab-4486-bee1-04cd24f7f430/5ace8d6f-49ab-4486-bee1-04cd24f7f430.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.150390] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-334d006d-dd49-4c18-b7b4-fda37262b339 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.158539] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 948.158539] env[62952]: value = "task-1367262" [ 948.158539] env[62952]: _type = "Task" [ 948.158539] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.168929] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367262, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.178955] env[62952]: DEBUG nova.network.neutron [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.220166] env[62952]: DEBUG nova.scheduler.client.report [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.234197] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367259, 'name': PowerOffVM_Task, 'duration_secs': 0.138378} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.234581] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.234842] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.235810] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b012b5b-ac19-4699-91eb-d9407d0d4059 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.243567] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.244649] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91c593e1-84f8-4e3b-86c2-1740e9dab96e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.271289] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.271621] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.271837] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Deleting the datastore file [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.272120] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5a2fa82-965c-4560-8f9f-34883e63d927 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.283967] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 948.283967] env[62952]: value = "task-1367264" [ 948.283967] env[62952]: _type = "Task" [ 948.283967] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.295192] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506968} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.296133] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] a88447f6-6a5a-490d-bf44-11c9237fa5cb/a88447f6-6a5a-490d-bf44-11c9237fa5cb.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.296571] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.296974] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d31acb3-710e-4fed-b854-00fca0a06c44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.305064] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367264, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.309847] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 948.309847] env[62952]: value = "task-1367265" [ 948.309847] env[62952]: _type = "Task" [ 948.309847] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.317824] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367265, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.433716] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367261, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.559141] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Releasing lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.559439] env[62952]: DEBUG nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Received event network-vif-plugged-4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.559640] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.559849] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.560019] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.560196] env[62952]: DEBUG nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] No waiting events found dispatching network-vif-plugged-4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 948.560366] env[62952]: WARNING nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Received unexpected event network-vif-plugged-4b40f3e8-7283-4edc-8016-d2a0bde62f60 for instance with vm_state building and task_state spawning. [ 948.560527] env[62952]: DEBUG nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Received event network-changed-4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.560708] env[62952]: DEBUG nova.compute.manager [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Refreshing instance network info cache due to event network-changed-4b40f3e8-7283-4edc-8016-d2a0bde62f60. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.560894] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.668486] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367262, 'name': Rename_Task, 'duration_secs': 0.172438} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.668817] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.669100] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bed4d601-1b9e-45e7-b171-a4d0ce1890b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.676465] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 948.676465] env[62952]: value = "task-1367266" [ 948.676465] env[62952]: _type = "Task" [ 948.676465] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.684566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.684845] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Instance network_info: |[{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.685136] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.685370] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.685555] env[62952]: DEBUG nova.network.neutron [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Refreshing network info cache for port 4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.686684] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:ab:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b40f3e8-7283-4edc-8016-d2a0bde62f60', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.695850] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating folder: Project (0c451895447b473eb18675f6bc843081). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 948.700216] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94af349c-4f60-4597-b867-2b8feb9ba03f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.714222] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created folder: Project (0c451895447b473eb18675f6bc843081) in parent group-v290852. [ 948.714222] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating folder: Instances. Parent ref: group-v290916. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 948.714222] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea2bd2f5-bae6-471d-8030-8c44c1033ec2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.726026] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created folder: Instances in parent group-v290916. [ 948.726026] env[62952]: DEBUG oslo.service.loopingcall [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.726026] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 948.726026] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f1d8d7f-0c56-4222-8d88-5fa514efd723 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.749025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.749025] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.753204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.357s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.754437] env[62952]: INFO nova.compute.claims [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.763474] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.763474] env[62952]: value = "task-1367269" [ 948.763474] env[62952]: _type = "Task" [ 948.763474] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.774731] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367269, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.795584] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367264, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092426} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.795886] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.796090] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.796273] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.828939] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367265, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072243} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.829300] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.830247] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3821eab2-ab76-4119-b097-a583ff162fc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.857831] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] a88447f6-6a5a-490d-bf44-11c9237fa5cb/a88447f6-6a5a-490d-bf44-11c9237fa5cb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.860153] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7ef0a73-b8d1-49fc-abf0-3ad4a1f9561f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.881385] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 948.881385] env[62952]: value = "task-1367270" [ 948.881385] env[62952]: _type = "Task" [ 948.881385] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.890288] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367270, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.936442] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367261, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.021854] env[62952]: DEBUG nova.network.neutron [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updated VIF entry in instance network info cache for port 4b40f3e8-7283-4edc-8016-d2a0bde62f60. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 949.022528] env[62952]: DEBUG nova.network.neutron [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.186940] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367266, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.195938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "82401449-347e-4917-a307-d5d7bc048f7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.196285] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "82401449-347e-4917-a307-d5d7bc048f7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.255528] env[62952]: DEBUG nova.compute.utils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.256556] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.256556] env[62952]: DEBUG nova.network.neutron [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.273789] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367269, 'name': CreateVM_Task, 'duration_secs': 0.410809} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.276169] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 949.276169] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.276169] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.276169] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.276169] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a63cca4-2547-45bb-b986-0e599600c778 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.280433] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 949.280433] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52246d96-5c76-fed7-3d60-3a67c3893217" [ 949.280433] env[62952]: _type = "Task" [ 949.280433] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.289720] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52246d96-5c76-fed7-3d60-3a67c3893217, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.307721] env[62952]: DEBUG nova.policy [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb541b4abd9426ea96edcd066c69be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d1f373267464f14ae2c833151821973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.394071] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367270, 'name': ReconfigVM_Task, 'duration_secs': 0.344967} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.394391] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Reconfigured VM instance instance-00000049 to attach disk [datastore1] a88447f6-6a5a-490d-bf44-11c9237fa5cb/a88447f6-6a5a-490d-bf44-11c9237fa5cb.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.395074] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df2f6dda-4017-487c-94f6-35da38fbfcb5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.402208] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 949.402208] env[62952]: value = "task-1367271" [ 949.402208] env[62952]: _type = "Task" [ 949.402208] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.411180] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367271, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.434138] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367261, 'name': CreateSnapshot_Task, 'duration_secs': 1.05532} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.434465] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 949.435272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d75510-a26d-4ebc-b0a8-d3db4a8d8c97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.521657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "97995e38-b27e-478a-8553-eb1c844bb0a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.521923] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "97995e38-b27e-478a-8553-eb1c844bb0a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.527288] env[62952]: DEBUG oslo_concurrency.lockutils [req-8b58efb4-9852-4d6d-b684-71892a48c3e8 req-dedbc3b7-cba1-4eb1-b26d-d90c791f17de service nova] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.589123] env[62952]: DEBUG nova.network.neutron [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Successfully created port: 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.692041] env[62952]: DEBUG oslo_vmware.api [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367266, 'name': PowerOnVM_Task, 'duration_secs': 0.738991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.692041] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.692041] env[62952]: INFO nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Took 8.09 seconds to spawn the instance on the hypervisor. [ 949.692041] env[62952]: DEBUG nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.692041] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc20da0-8655-4420-b845-21cd4164d3a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.698689] env[62952]: DEBUG nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 949.764017] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.793125] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52246d96-5c76-fed7-3d60-3a67c3893217, 'name': SearchDatastore_Task, 'duration_secs': 0.009269} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.793125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.793125] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.793125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.793473] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.793473] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.793473] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e399828-3934-4650-9b11-c4abc0284a85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.803269] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 949.803269] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 949.805375] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac47744f-89d5-4051-b76a-1efb79bd38da {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.814060] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 949.814060] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528268eb-80b8-13fc-31e4-dfe51ad3080c" [ 949.814060] env[62952]: _type = "Task" [ 949.814060] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.822592] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528268eb-80b8-13fc-31e4-dfe51ad3080c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.849105] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.849105] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.849105] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.849105] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.849278] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.849278] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.849278] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.849278] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.849278] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.849424] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.849424] env[62952]: DEBUG nova.virt.hardware [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.850596] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8de6fc2-7b28-4fab-b9ce-207045f6b8b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.861744] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e9cbef-2d78-46cb-85c1-b92c34443d5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.876471] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.882144] env[62952]: DEBUG oslo.service.loopingcall [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.884991] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.885390] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9decbbf4-bef9-4a91-b63d-bf2876b7b8e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.907836] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.907836] env[62952]: value = "task-1367272" [ 949.907836] env[62952]: _type = "Task" [ 949.907836] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.914064] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367271, 'name': Rename_Task, 'duration_secs': 0.131526} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.914678] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.914975] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f31ee32-85ff-471e-b9df-1ea91cfa254e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.921601] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367272, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.927118] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 949.927118] env[62952]: value = "task-1367273" [ 949.927118] env[62952]: _type = "Task" [ 949.927118] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.934599] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367273, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.955488] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 949.958780] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7f3fe6e0-26be-4b15-81c8-81139977447c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.967089] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 949.967089] env[62952]: value = "task-1367274" [ 949.967089] env[62952]: _type = "Task" [ 949.967089] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.978292] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367274, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.025614] env[62952]: DEBUG nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.113982] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be9aa94-7435-401c-a52c-f2a24f90fe53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.121393] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f4855e-9521-4cf7-b62f-702a36ff72a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.154460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f0a85b-f11c-4d76-951d-6ba5cb535bb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.164029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa9db40-2e8f-46d9-ac88-5832c3604111 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.181061] env[62952]: DEBUG nova.compute.provider_tree [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.210940] env[62952]: INFO nova.compute.manager [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Took 27.74 seconds to build instance. [ 950.228738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.326035] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528268eb-80b8-13fc-31e4-dfe51ad3080c, 'name': SearchDatastore_Task, 'duration_secs': 0.008221} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.327036] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15c07e1c-e3df-45bd-8bce-1f80b785b14d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.332714] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 950.332714] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b42537-cbe6-b2d1-d1ab-ef84eb470ab1" [ 950.332714] env[62952]: _type = "Task" [ 950.332714] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.341218] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b42537-cbe6-b2d1-d1ab-ef84eb470ab1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.422497] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367272, 'name': CreateVM_Task, 'duration_secs': 0.488922} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.422780] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.423644] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.423938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.424399] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.424769] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8550a92a-7110-498b-8b88-4eadd304c763 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.433768] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 950.433768] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5270e8ff-8d0b-7087-67b5-118265b1f125" [ 950.433768] env[62952]: _type = "Task" [ 950.433768] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.443752] env[62952]: DEBUG oslo_vmware.api [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367273, 'name': PowerOnVM_Task, 'duration_secs': 0.466223} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.444655] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 950.445013] env[62952]: INFO nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Took 6.43 seconds to spawn the instance on the hypervisor. [ 950.445335] env[62952]: DEBUG nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.446512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff94094-f415-4424-8ea6-21ff91f9cc55 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.455650] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5270e8ff-8d0b-7087-67b5-118265b1f125, 'name': SearchDatastore_Task, 'duration_secs': 0.010234} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.456453] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.456806] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.457226] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.457445] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.457810] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.459906] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49078ce3-951b-48c9-aded-94f950b6a3cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.473468] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.473595] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.474997] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f134948-637d-475b-ab25-2beef26e4502 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.483033] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367274, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.486472] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 950.486472] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52dd0e4d-9914-04ce-d393-d24dc629d7ad" [ 950.486472] env[62952]: _type = "Task" [ 950.486472] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.497902] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52dd0e4d-9914-04ce-d393-d24dc629d7ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.545686] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.685807] env[62952]: DEBUG nova.scheduler.client.report [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.714975] env[62952]: DEBUG oslo_concurrency.lockutils [None req-360ed9e2-06e4-46ce-bc22-0e7bdff8ff2a tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.507s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.779257] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.801509] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.801786] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.801963] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.802338] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.802537] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.802725] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.802945] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.803122] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.803294] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.803455] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.803626] env[62952]: DEBUG nova.virt.hardware [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.804494] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45bd6c5-a309-4fc6-b971-0c96dfcf6000 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.812206] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b89b76-d59a-4355-b369-f10dfba9f0e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.840441] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b42537-cbe6-b2d1-d1ab-ef84eb470ab1, 'name': SearchDatastore_Task, 'duration_secs': 0.00987} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.840670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.840918] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2/867ffed2-50ec-45d8-b64e-989cf8d1b0e2.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 950.841156] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f19a4cfa-9f2d-41b5-b775-d741aab27264 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.847454] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 950.847454] env[62952]: value = "task-1367275" [ 950.847454] env[62952]: _type = "Task" [ 950.847454] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.854829] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367275, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.905531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.906044] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.906331] env[62952]: INFO nova.compute.manager [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Shelving [ 950.979834] env[62952]: INFO nova.compute.manager [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Took 25.57 seconds to build instance. [ 950.985631] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367274, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.002056] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52dd0e4d-9914-04ce-d393-d24dc629d7ad, 'name': SearchDatastore_Task, 'duration_secs': 0.008585} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.002970] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-610edd21-0cd8-4ea1-8ece-1d89f9f6326b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.009238] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 951.009238] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52361ab9-06ac-fca2-b050-7e09fff413dd" [ 951.009238] env[62952]: _type = "Task" [ 951.009238] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.017892] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52361ab9-06ac-fca2-b050-7e09fff413dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.149696] env[62952]: DEBUG nova.compute.manager [req-9fe3393d-1db9-4239-9b91-45d8aad8064d req-7410864f-ddd9-4cfd-97ad-cb0faee38aee service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-vif-plugged-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.149945] env[62952]: DEBUG oslo_concurrency.lockutils [req-9fe3393d-1db9-4239-9b91-45d8aad8064d req-7410864f-ddd9-4cfd-97ad-cb0faee38aee service nova] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.150165] env[62952]: DEBUG oslo_concurrency.lockutils [req-9fe3393d-1db9-4239-9b91-45d8aad8064d req-7410864f-ddd9-4cfd-97ad-cb0faee38aee service nova] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.150332] env[62952]: DEBUG oslo_concurrency.lockutils [req-9fe3393d-1db9-4239-9b91-45d8aad8064d req-7410864f-ddd9-4cfd-97ad-cb0faee38aee service nova] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.150493] env[62952]: DEBUG nova.compute.manager [req-9fe3393d-1db9-4239-9b91-45d8aad8064d req-7410864f-ddd9-4cfd-97ad-cb0faee38aee service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] No waiting events found dispatching network-vif-plugged-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.150653] env[62952]: WARNING nova.compute.manager [req-9fe3393d-1db9-4239-9b91-45d8aad8064d req-7410864f-ddd9-4cfd-97ad-cb0faee38aee service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received unexpected event network-vif-plugged-48403971-25e5-4da0-a350-bd47ba090afc for instance with vm_state building and task_state spawning. [ 951.191201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.191731] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.194373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.267s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.194677] env[62952]: DEBUG nova.objects.instance [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lazy-loading 'resources' on Instance uuid 2f475040-28df-47e4-bfc4-4c843a57885f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.296619] env[62952]: DEBUG nova.network.neutron [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Successfully updated port: 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.356897] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367275, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481927} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.357472] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2/867ffed2-50ec-45d8-b64e-989cf8d1b0e2.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.357472] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.357645] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-897d51e1-ff0c-4758-a670-ba4d096655de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.364958] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 951.364958] env[62952]: value = "task-1367276" [ 951.364958] env[62952]: _type = "Task" [ 951.364958] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.373117] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367276, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.414061] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.414342] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb5f8402-ecbb-4317-874c-bd80117b8810 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.421084] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 951.421084] env[62952]: value = "task-1367277" [ 951.421084] env[62952]: _type = "Task" [ 951.421084] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.431429] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.482471] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367274, 'name': CloneVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.487165] env[62952]: DEBUG oslo_concurrency.lockutils [None req-57d817f2-77ac-4908-b85f-93c5d20238d0 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.657s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.521036] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52361ab9-06ac-fca2-b050-7e09fff413dd, 'name': SearchDatastore_Task, 'duration_secs': 0.014284} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.521355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.521626] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.522053] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c04c26ba-94fa-4898-9ee1-e9b1ef9dbec3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.528801] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 951.528801] env[62952]: value = "task-1367278" [ 951.528801] env[62952]: _type = "Task" [ 951.528801] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.536568] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.698133] env[62952]: DEBUG nova.compute.utils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.703531] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.703759] env[62952]: DEBUG nova.network.neutron [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.719661] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.720119] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.720400] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.720594] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.720760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.723471] env[62952]: INFO nova.compute.manager [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Terminating instance [ 951.725389] env[62952]: DEBUG nova.compute.manager [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 951.726037] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 951.726485] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8334c0-d171-4027-9db3-370dd1a7d806 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.736264] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.736748] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bb48d51-e45d-443c-b758-68bb091159bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.743560] env[62952]: DEBUG oslo_vmware.api [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 951.743560] env[62952]: value = "task-1367279" [ 951.743560] env[62952]: _type = "Task" [ 951.743560] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.754362] env[62952]: DEBUG oslo_vmware.api [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.755960] env[62952]: DEBUG nova.policy [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.801612] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.802029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.802029] env[62952]: DEBUG nova.network.neutron [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.878993] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367276, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096581} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.881501] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.882779] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f86f216-785e-4dec-838d-dddfe69f3bb6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.908340] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2/867ffed2-50ec-45d8-b64e-989cf8d1b0e2.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.911301] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c2e0aaa-e40f-4f51-85bf-041195ac4c1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.937729] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367277, 'name': PowerOffVM_Task, 'duration_secs': 0.217008} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.938731] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.939279] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 951.939279] env[62952]: value = "task-1367280" [ 951.939279] env[62952]: _type = "Task" [ 951.939279] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.939913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3581493-39f5-4b90-8eef-17196ef39571 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.954428] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367280, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.973479] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb858a9-052d-43c7-9e47-9448c0d6a411 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.988884] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367274, 'name': CloneVM_Task, 'duration_secs': 1.575965} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.989389] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Created linked-clone VM from snapshot [ 951.990132] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a15bb-ed0a-4ef1-88e8-31282df761f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.999012] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Uploading image 1449561c-f27a-4b93-87d2-4669ae19d131 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 952.041145] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46847} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.043291] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 952.043291] env[62952]: value = "vm-290921" [ 952.043291] env[62952]: _type = "VirtualMachine" [ 952.043291] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 952.043587] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.043795] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.044034] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cb72775d-9832-44b9-989b-207ede561bde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.045825] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6947c448-91ee-4db4-b9be-f741b372b15a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.055919] env[62952]: DEBUG nova.network.neutron [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Successfully created port: be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.057957] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lease: (returnval){ [ 952.057957] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526e815f-8cc7-64ce-d351-021cc857d190" [ 952.057957] env[62952]: _type = "HttpNfcLease" [ 952.057957] env[62952]: } obtained for exporting VM: (result){ [ 952.057957] env[62952]: value = "vm-290921" [ 952.057957] env[62952]: _type = "VirtualMachine" [ 952.057957] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 952.057957] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the lease: (returnval){ [ 952.057957] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526e815f-8cc7-64ce-d351-021cc857d190" [ 952.057957] env[62952]: _type = "HttpNfcLease" [ 952.057957] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 952.059692] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 952.059692] env[62952]: value = "task-1367282" [ 952.059692] env[62952]: _type = "Task" [ 952.059692] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.075717] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 952.075717] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526e815f-8cc7-64ce-d351-021cc857d190" [ 952.075717] env[62952]: _type = "HttpNfcLease" [ 952.075717] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 952.075717] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367282, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.101589] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31ddeda-01db-418b-8c09-816d35d8884a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.109170] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac17ee7f-4365-4450-a5bd-9b49352cfd95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.140673] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc7629f-c02f-4715-98eb-f12bce62759b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.148489] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf853f4e-be1a-4c71-8769-9d632414f28b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.161473] env[62952]: DEBUG nova.compute.provider_tree [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.203550] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.254762] env[62952]: DEBUG oslo_vmware.api [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367279, 'name': PowerOffVM_Task, 'duration_secs': 0.182076} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.255193] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 952.255397] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 952.255647] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4083088e-7acc-4741-859c-6551959f5793 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.338213] env[62952]: DEBUG nova.network.neutron [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.357980] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 952.358217] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 952.359608] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Deleting the datastore file [datastore1] a88447f6-6a5a-490d-bf44-11c9237fa5cb {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.359608] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-752d3ecb-be89-46e6-a680-d2154cf5ec2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.364929] env[62952]: DEBUG oslo_vmware.api [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for the task: (returnval){ [ 952.364929] env[62952]: value = "task-1367284" [ 952.364929] env[62952]: _type = "Task" [ 952.364929] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.372832] env[62952]: DEBUG oslo_vmware.api [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367284, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.453564] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367280, 'name': ReconfigVM_Task, 'duration_secs': 0.283594} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.453848] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2/867ffed2-50ec-45d8-b64e-989cf8d1b0e2.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.454557] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-377332ac-bc2a-4e71-9c4c-c6bf75e4659f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.460759] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 952.460759] env[62952]: value = "task-1367285" [ 952.460759] env[62952]: _type = "Task" [ 952.460759] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.469057] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367285, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.491274] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 952.491660] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d722c377-5761-48da-9838-9a5bc690dea8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.498689] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 952.498689] env[62952]: value = "task-1367286" [ 952.498689] env[62952]: _type = "Task" [ 952.498689] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.506954] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367286, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.532462] env[62952]: DEBUG nova.network.neutron [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.570127] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 952.570127] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526e815f-8cc7-64ce-d351-021cc857d190" [ 952.570127] env[62952]: _type = "HttpNfcLease" [ 952.570127] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 952.574077] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 952.574077] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526e815f-8cc7-64ce-d351-021cc857d190" [ 952.574077] env[62952]: _type = "HttpNfcLease" [ 952.574077] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 952.574447] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367282, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069891} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.575229] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9243b276-91c8-4dd5-8765-c81c101bcaea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.578038] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.579080] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc6bdde-652b-428b-baa8-d3ee16ae083c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.586805] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f776ef-90e9-ddf6-aa4a-74f1e61ea052/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 952.587012] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f776ef-90e9-ddf6-aa4a-74f1e61ea052/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 952.604372] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.605831] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f4a623d-6e34-48f4-9ecd-33c5f2df5867 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.675856] env[62952]: DEBUG nova.scheduler.client.report [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.684734] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 952.684734] env[62952]: value = "task-1367287" [ 952.684734] env[62952]: _type = "Task" [ 952.684734] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.693657] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367287, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.740438] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c302aff1-a7cc-471a-9a96-7e4c0b0bcb93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.873894] env[62952]: DEBUG oslo_vmware.api [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Task: {'id': task-1367284, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179892} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.874215] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.874429] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 952.874609] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 952.874929] env[62952]: INFO nova.compute.manager [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Took 1.15 seconds to destroy the instance on the hypervisor. [ 952.875259] env[62952]: DEBUG oslo.service.loopingcall [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.875460] env[62952]: DEBUG nova.compute.manager [-] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 952.875556] env[62952]: DEBUG nova.network.neutron [-] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 952.971803] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367285, 'name': Rename_Task, 'duration_secs': 0.153797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.973591] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.973591] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8757721e-7203-4692-aaaa-8c07cde22109 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.980573] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 952.980573] env[62952]: value = "task-1367288" [ 952.980573] env[62952]: _type = "Task" [ 952.980573] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.989127] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367288, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.008827] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367286, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.035127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.035435] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Instance network_info: |[{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 953.036026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:99:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48403971-25e5-4da0-a350-bd47ba090afc', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.045281] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Creating folder: Project (5d1f373267464f14ae2c833151821973). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 953.045673] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-80c77657-bcda-4f34-909d-0856f2470973 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.057072] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Created folder: Project (5d1f373267464f14ae2c833151821973) in parent group-v290852. [ 953.057303] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Creating folder: Instances. Parent ref: group-v290922. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 953.057563] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1bb52ec-efce-43f0-ab8e-8a55a624f7b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.067851] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Created folder: Instances in parent group-v290922. [ 953.069027] env[62952]: DEBUG oslo.service.loopingcall [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.069027] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.069027] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5861af60-0727-495b-aa48-31571e84ea4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.091511] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.091511] env[62952]: value = "task-1367291" [ 953.091511] env[62952]: _type = "Task" [ 953.091511] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.099836] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367291, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.169129] env[62952]: DEBUG nova.compute.manager [req-bde1fecd-9d5d-4a51-81dc-e804eaa0ddc3 req-64affa1e-001d-4c3b-bd9a-8b888f586484 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Received event network-vif-deleted-104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.169345] env[62952]: INFO nova.compute.manager [req-bde1fecd-9d5d-4a51-81dc-e804eaa0ddc3 req-64affa1e-001d-4c3b-bd9a-8b888f586484 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Neutron deleted interface 104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2; detaching it from the instance and deleting it from the info cache [ 953.171045] env[62952]: DEBUG nova.network.neutron [req-bde1fecd-9d5d-4a51-81dc-e804eaa0ddc3 req-64affa1e-001d-4c3b-bd9a-8b888f586484 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.182558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.188201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.008s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.188201] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.188201] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 953.188384] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.729s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.191218] env[62952]: INFO nova.compute.claims [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.199957] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a50f312-28a6-4c83-a0e6-735f4cc058f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.209936] env[62952]: DEBUG nova.compute.manager [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-changed-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.210439] env[62952]: DEBUG nova.compute.manager [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing instance network info cache due to event network-changed-48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.211124] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.211320] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.211642] env[62952]: DEBUG nova.network.neutron [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.214851] env[62952]: INFO nova.scheduler.client.report [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleted allocations for instance 2f475040-28df-47e4-bfc4-4c843a57885f [ 953.217551] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.230032] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367287, 'name': ReconfigVM_Task, 'duration_secs': 0.515778} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.230593] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Reconfigured VM instance instance-00000047 to attach disk [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02/a3c01d4d-9c66-4441-95e8-87998782cc02.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.231640] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00b5c5eb-3ed0-4238-90d2-e2f8c92549c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.234414] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028269be-a133-48f0-beb8-992ebf03867d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.255782] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcb471e-6e92-4060-80b2-bb799a823fb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.260144] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 953.260144] env[62952]: value = "task-1367292" [ 953.260144] env[62952]: _type = "Task" [ 953.260144] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.267645] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.267910] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.268168] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.268552] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.268688] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.268853] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.269142] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.269448] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.269660] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.269966] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.270068] env[62952]: DEBUG nova.virt.hardware [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.273399] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753fcf9f-486a-4a04-b7ea-1f973023753c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.277683] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de9648e-d071-4911-a1b9-6c4cc4d0487d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.285540] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367292, 'name': Rename_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.315279] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b777ff34-33c3-43e2-9410-2a92166bad12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.320966] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181437MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 953.321164] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.492212] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367288, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.509354] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367286, 'name': CreateSnapshot_Task, 'duration_secs': 0.868721} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.510095] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 953.511021] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2dca94-91ca-427a-af27-f6329dfad576 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.596815] env[62952]: DEBUG nova.network.neutron [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Successfully updated port: be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.606522] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367291, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.613113] env[62952]: DEBUG nova.network.neutron [-] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.675225] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df490349-2cf1-4c61-b69a-e0d7e7079f12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.684325] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf62d5b-3e43-4564-a84c-2b9e2c2f740c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.713173] env[62952]: DEBUG nova.compute.manager [req-bde1fecd-9d5d-4a51-81dc-e804eaa0ddc3 req-64affa1e-001d-4c3b-bd9a-8b888f586484 service nova] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Detach interface failed, port_id=104e8b8a-aaaa-4d4e-9f58-c5eeaf5aa4d2, reason: Instance a88447f6-6a5a-490d-bf44-11c9237fa5cb could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 953.729369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8338da47-7d30-49e1-bb68-fe6dfc189a7d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "2f475040-28df-47e4-bfc4-4c843a57885f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.058s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.776764] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367292, 'name': Rename_Task, 'duration_secs': 0.158752} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.777348] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.777770] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e18fec7-64e8-476f-816a-215964ef2442 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.791701] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 953.791701] env[62952]: value = "task-1367293" [ 953.791701] env[62952]: _type = "Task" [ 953.791701] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.807631] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367293, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.000422] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367288, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.029226] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 954.029694] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9ac2f7b5-f830-4ecd-980d-d6fbd9044ba9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.041038] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 954.041038] env[62952]: value = "task-1367294" [ 954.041038] env[62952]: _type = "Task" [ 954.041038] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.048343] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367294, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.064375] env[62952]: DEBUG nova.network.neutron [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updated VIF entry in instance network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.065011] env[62952]: DEBUG nova.network.neutron [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.100035] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367291, 'name': CreateVM_Task, 'duration_secs': 0.837483} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.100035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-532a2bcd-5d6a-4fa9-abc9-a048cc915fda" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.100035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-532a2bcd-5d6a-4fa9-abc9-a048cc915fda" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.100035] env[62952]: DEBUG nova.network.neutron [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.101016] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.101670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.101847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.102194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.102648] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3c70939-f0b4-4414-addf-7a042a92fd4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.107965] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 954.107965] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52de499e-9fc0-5092-ca9e-4a28ec597e7e" [ 954.107965] env[62952]: _type = "Task" [ 954.107965] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.118545] env[62952]: INFO nova.compute.manager [-] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Took 1.24 seconds to deallocate network for instance. [ 954.119116] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52de499e-9fc0-5092-ca9e-4a28ec597e7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.303128] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367293, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.490928] env[62952]: DEBUG oslo_vmware.api [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367288, 'name': PowerOnVM_Task, 'duration_secs': 1.390046} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.493495] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.493722] env[62952]: INFO nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 8.01 seconds to spawn the instance on the hypervisor. [ 954.493905] env[62952]: DEBUG nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.495018] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b89c60-4c5e-4eea-871d-1add3f8984d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.510127] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83dc083-e0c6-495f-bbe7-d435cc102553 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.516742] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad6632e-d00a-4dd2-87a1-089bd495e7fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.551194] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaafbeb2-6bb8-45a4-8fc9-40cc479df6ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.562492] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a831f765-8bf6-4251-8fce-5100d573ffdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.566486] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367294, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.567219] env[62952]: DEBUG oslo_concurrency.lockutils [req-7a0e9a59-82b3-4dec-9030-f691c6ed2510 req-6ccb8211-f46b-442f-a7c4-56296cfb66b0 service nova] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.577738] env[62952]: DEBUG nova.compute.provider_tree [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.617368] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52de499e-9fc0-5092-ca9e-4a28ec597e7e, 'name': SearchDatastore_Task, 'duration_secs': 0.016525} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.617669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.617906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.618163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.618311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.618489] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.619039] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc99477d-8073-446c-aa68-103e1655cde0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.624925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.627426] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.627604] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.628319] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36772ac9-d476-48ec-b410-b09c8a68b2c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.633649] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 954.633649] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523df0d0-3689-14c5-0c4a-9236d454d6a6" [ 954.633649] env[62952]: _type = "Task" [ 954.633649] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.634376] env[62952]: DEBUG nova.network.neutron [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.646107] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523df0d0-3689-14c5-0c4a-9236d454d6a6, 'name': SearchDatastore_Task, 'duration_secs': 0.009192} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.646678] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7571f40-67b9-4f6e-a0a6-ecaf0ec30f8e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.651746] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 954.651746] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52235b4f-5878-65a7-bef0-68dd9a65c875" [ 954.651746] env[62952]: _type = "Task" [ 954.651746] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.661347] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52235b4f-5878-65a7-bef0-68dd9a65c875, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.792873] env[62952]: DEBUG nova.network.neutron [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Updating instance_info_cache with network_info: [{"id": "be66392d-0f28-4019-9db6-f989ba14dbe6", "address": "fa:16:3e:43:19:9c", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe66392d-0f", "ovs_interfaceid": "be66392d-0f28-4019-9db6-f989ba14dbe6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.804229] env[62952]: DEBUG oslo_vmware.api [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367293, 'name': PowerOnVM_Task, 'duration_secs': 0.74223} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.804537] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.804855] env[62952]: DEBUG nova.compute.manager [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.805634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6967cc-c1ba-45af-9b59-fa2cfa782443 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.814610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "802fd770-38aa-4417-b25e-fd83510bceef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.814882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "802fd770-38aa-4417-b25e-fd83510bceef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.815105] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "802fd770-38aa-4417-b25e-fd83510bceef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.815402] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "802fd770-38aa-4417-b25e-fd83510bceef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.815497] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "802fd770-38aa-4417-b25e-fd83510bceef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.818817] env[62952]: INFO nova.compute.manager [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Terminating instance [ 954.820644] env[62952]: DEBUG nova.compute.manager [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.820873] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.821729] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fd0da6-cbc1-4324-bfcf-3ce5cdb05c47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.830091] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.830091] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04acc75d-c697-4f21-81e7-5608083fa9f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.836059] env[62952]: DEBUG oslo_vmware.api [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 954.836059] env[62952]: value = "task-1367295" [ 954.836059] env[62952]: _type = "Task" [ 954.836059] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.844321] env[62952]: DEBUG oslo_vmware.api [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367295, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.021046] env[62952]: INFO nova.compute.manager [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 27.57 seconds to build instance. [ 955.058183] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367294, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.080901] env[62952]: DEBUG nova.scheduler.client.report [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.162432] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52235b4f-5878-65a7-bef0-68dd9a65c875, 'name': SearchDatastore_Task, 'duration_secs': 0.010776} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.162717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.163038] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] ff189c9e-9e7c-4217-9c65-0f821393870a/ff189c9e-9e7c-4217-9c65-0f821393870a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.163328] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70b355ab-7d8a-44e4-a635-2bea57e1af15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.172348] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 955.172348] env[62952]: value = "task-1367296" [ 955.172348] env[62952]: _type = "Task" [ 955.172348] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.181847] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367296, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.258435] env[62952]: DEBUG nova.compute.manager [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Received event network-vif-plugged-be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.258702] env[62952]: DEBUG oslo_concurrency.lockutils [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] Acquiring lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.259027] env[62952]: DEBUG oslo_concurrency.lockutils [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.259089] env[62952]: DEBUG oslo_concurrency.lockutils [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.259284] env[62952]: DEBUG nova.compute.manager [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] No waiting events found dispatching network-vif-plugged-be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.259608] env[62952]: WARNING nova.compute.manager [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Received unexpected event network-vif-plugged-be66392d-0f28-4019-9db6-f989ba14dbe6 for instance with vm_state building and task_state spawning. [ 955.259716] env[62952]: DEBUG nova.compute.manager [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Received event network-changed-be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.259841] env[62952]: DEBUG nova.compute.manager [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Refreshing instance network info cache due to event network-changed-be66392d-0f28-4019-9db6-f989ba14dbe6. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.259978] env[62952]: DEBUG oslo_concurrency.lockutils [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] Acquiring lock "refresh_cache-532a2bcd-5d6a-4fa9-abc9-a048cc915fda" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.298815] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-532a2bcd-5d6a-4fa9-abc9-a048cc915fda" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.299054] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Instance network_info: |[{"id": "be66392d-0f28-4019-9db6-f989ba14dbe6", "address": "fa:16:3e:43:19:9c", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe66392d-0f", "ovs_interfaceid": "be66392d-0f28-4019-9db6-f989ba14dbe6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.299408] env[62952]: DEBUG oslo_concurrency.lockutils [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] Acquired lock "refresh_cache-532a2bcd-5d6a-4fa9-abc9-a048cc915fda" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.299602] env[62952]: DEBUG nova.network.neutron [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Refreshing network info cache for port be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.301108] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:19:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be66392d-0f28-4019-9db6-f989ba14dbe6', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.309236] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating folder: Project (30b8e72818064a0f950614b39e30f717). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 955.309998] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33e1b5c8-8bee-4647-8a70-ff2803c22cf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.326180] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created folder: Project (30b8e72818064a0f950614b39e30f717) in parent group-v290852. [ 955.326180] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating folder: Instances. Parent ref: group-v290927. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 955.326180] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.326498] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bccfcc8-228c-4063-9a1f-9e1ebe114d9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.335892] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created folder: Instances in parent group-v290927. [ 955.336225] env[62952]: DEBUG oslo.service.loopingcall [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.336446] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.336796] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6196f22-a340-4ad3-80b6-76780601a7c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.366120] env[62952]: DEBUG oslo_vmware.api [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367295, 'name': PowerOffVM_Task, 'duration_secs': 0.246632} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.367971] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.368243] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.368563] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.368563] env[62952]: value = "task-1367299" [ 955.368563] env[62952]: _type = "Task" [ 955.368563] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.368881] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8fca044-b31a-40c3-82e9-16f5b376a293 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.379744] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367299, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.531394] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5eb9ad2d-3c3d-498b-98d1-8932f38b89a6 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.024s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.550425] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.550754] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.550952] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleting the datastore file [datastore2] 802fd770-38aa-4417-b25e-fd83510bceef {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.556821] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16c7c8d6-60e5-4af7-86a3-37b83b1862cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.571941] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367294, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.575191] env[62952]: DEBUG oslo_vmware.api [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 955.575191] env[62952]: value = "task-1367301" [ 955.575191] env[62952]: _type = "Task" [ 955.575191] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.591162] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.591945] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.597434] env[62952]: DEBUG oslo_vmware.api [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.597893] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.370s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.598279] env[62952]: DEBUG nova.objects.instance [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lazy-loading 'resources' on Instance uuid 4983f354-77dc-4e84-969b-4c74441fa568 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.683400] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367296, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.884058] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367299, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.920546] env[62952]: DEBUG nova.compute.manager [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Received event network-changed-4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.920743] env[62952]: DEBUG nova.compute.manager [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Refreshing instance network info cache due to event network-changed-4b40f3e8-7283-4edc-8016-d2a0bde62f60. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.920926] env[62952]: DEBUG oslo_concurrency.lockutils [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.921121] env[62952]: DEBUG oslo_concurrency.lockutils [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.921233] env[62952]: DEBUG nova.network.neutron [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Refreshing network info cache for port 4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 956.012658] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "a3c01d4d-9c66-4441-95e8-87998782cc02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.012936] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "a3c01d4d-9c66-4441-95e8-87998782cc02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.013259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "a3c01d4d-9c66-4441-95e8-87998782cc02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.013390] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "a3c01d4d-9c66-4441-95e8-87998782cc02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.013684] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "a3c01d4d-9c66-4441-95e8-87998782cc02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.015770] env[62952]: INFO nova.compute.manager [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Terminating instance [ 956.017475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "refresh_cache-a3c01d4d-9c66-4441-95e8-87998782cc02" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.017638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquired lock "refresh_cache-a3c01d4d-9c66-4441-95e8-87998782cc02" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.017903] env[62952]: DEBUG nova.network.neutron [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.032889] env[62952]: DEBUG nova.network.neutron [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Updated VIF entry in instance network info cache for port be66392d-0f28-4019-9db6-f989ba14dbe6. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.033310] env[62952]: DEBUG nova.network.neutron [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Updating instance_info_cache with network_info: [{"id": "be66392d-0f28-4019-9db6-f989ba14dbe6", "address": "fa:16:3e:43:19:9c", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe66392d-0f", "ovs_interfaceid": "be66392d-0f28-4019-9db6-f989ba14dbe6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.059816] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367294, 'name': CloneVM_Task, 'duration_secs': 1.714721} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.060094] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Created linked-clone VM from snapshot [ 956.060845] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6a1170-296c-44e3-b3e5-9d2d5b6b0274 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.068467] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Uploading image f9471984-046d-4570-b0fc-4791884b7f4f {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 956.085230] env[62952]: DEBUG oslo_vmware.api [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339006} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.087252] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.087449] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.087628] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.087809] env[62952]: INFO nova.compute.manager [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Took 1.27 seconds to destroy the instance on the hypervisor. [ 956.088092] env[62952]: DEBUG oslo.service.loopingcall [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.089631] env[62952]: DEBUG nova.compute.manager [-] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.089631] env[62952]: DEBUG nova.network.neutron [-] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.093787] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 956.093787] env[62952]: value = "vm-290926" [ 956.093787] env[62952]: _type = "VirtualMachine" [ 956.093787] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 956.094037] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c4d3be84-6fcd-4106-9ea0-89b3ba26334b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.098842] env[62952]: DEBUG nova.compute.utils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.101635] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.101809] env[62952]: DEBUG nova.network.neutron [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 956.106494] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lease: (returnval){ [ 956.106494] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5230a09b-43a6-97ed-9226-12db3b7436a1" [ 956.106494] env[62952]: _type = "HttpNfcLease" [ 956.106494] env[62952]: } obtained for exporting VM: (result){ [ 956.106494] env[62952]: value = "vm-290926" [ 956.106494] env[62952]: _type = "VirtualMachine" [ 956.106494] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 956.107214] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the lease: (returnval){ [ 956.107214] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5230a09b-43a6-97ed-9226-12db3b7436a1" [ 956.107214] env[62952]: _type = "HttpNfcLease" [ 956.107214] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 956.114666] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.114666] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5230a09b-43a6-97ed-9226-12db3b7436a1" [ 956.114666] env[62952]: _type = "HttpNfcLease" [ 956.114666] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 956.150884] env[62952]: DEBUG nova.policy [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2dd3ee23972e46178d446f81de42acdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1396825e1044402b7d1c799657c1b94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 956.187056] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367296, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586422} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.187467] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] ff189c9e-9e7c-4217-9c65-0f821393870a/ff189c9e-9e7c-4217-9c65-0f821393870a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.187739] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.188060] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1824938-0cab-4f74-a61c-2cdb59366dbf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.196121] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 956.196121] env[62952]: value = "task-1367303" [ 956.196121] env[62952]: _type = "Task" [ 956.196121] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.208658] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.383358] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367299, 'name': CreateVM_Task, 'duration_secs': 0.542808} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.383358] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.384051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.384229] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.384613] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.384873] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a150206-4e37-4775-a41f-37f9907ce56a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.389490] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 956.389490] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52da82e8-370a-0816-8788-f0f929a8040b" [ 956.389490] env[62952]: _type = "Task" [ 956.389490] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.398978] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52da82e8-370a-0816-8788-f0f929a8040b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.447909] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36052748-bfc9-43d8-ab87-557b9dab087f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.455674] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7481e7-0c95-4ed4-813e-de317053933c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.488354] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91d9644-c3b7-48d6-8bbb-a2619baa876f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.496082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be75dd87-c3c6-4ec1-9a14-da070a912661 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.513135] env[62952]: DEBUG nova.compute.provider_tree [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.537489] env[62952]: DEBUG oslo_concurrency.lockutils [req-860f1d09-4b97-4724-8ae7-318f1b6b0dee req-029b4eb3-9350-4360-9658-e5c21c46c53c service nova] Releasing lock "refresh_cache-532a2bcd-5d6a-4fa9-abc9-a048cc915fda" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.547406] env[62952]: DEBUG nova.network.neutron [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Successfully created port: 178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.549886] env[62952]: DEBUG nova.network.neutron [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.608679] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.632690] env[62952]: DEBUG nova.network.neutron [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.634030] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.634030] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5230a09b-43a6-97ed-9226-12db3b7436a1" [ 956.634030] env[62952]: _type = "HttpNfcLease" [ 956.634030] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 956.634935] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 956.634935] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5230a09b-43a6-97ed-9226-12db3b7436a1" [ 956.634935] env[62952]: _type = "HttpNfcLease" [ 956.634935] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 956.636031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcfd598-8cf2-4781-b08f-2b41788faedb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.645821] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecc5f7-757f-40ff-6ebd-450fc40574d3/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 956.646137] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecc5f7-757f-40ff-6ebd-450fc40574d3/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 956.719570] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080486} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.719936] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.720828] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942cff29-cbe6-4c07-a4d6-bce15168e1a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.749788] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] ff189c9e-9e7c-4217-9c65-0f821393870a/ff189c9e-9e7c-4217-9c65-0f821393870a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.750184] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c57eb48-94f7-4676-9026-c43988a19f0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.773229] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2ad239b7-a5be-452d-9e81-9fd770b2de81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.778243] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 956.778243] env[62952]: value = "task-1367304" [ 956.778243] env[62952]: _type = "Task" [ 956.778243] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.788162] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367304, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.840542] env[62952]: DEBUG nova.network.neutron [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updated VIF entry in instance network info cache for port 4b40f3e8-7283-4edc-8016-d2a0bde62f60. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.840542] env[62952]: DEBUG nova.network.neutron [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.860973] env[62952]: DEBUG nova.network.neutron [-] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.900511] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52da82e8-370a-0816-8788-f0f929a8040b, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.900908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.901226] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.901504] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.901664] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.901857] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.902154] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a18d171-6f1a-48cf-897d-cc142af6ebe8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.910946] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.911301] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.912646] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa612488-683a-44be-a2b7-763770dd0914 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.919325] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 956.919325] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523fa0a6-0952-7fef-0ae5-7daa0d526700" [ 956.919325] env[62952]: _type = "Task" [ 956.919325] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.927523] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523fa0a6-0952-7fef-0ae5-7daa0d526700, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.019364] env[62952]: DEBUG nova.scheduler.client.report [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.136994] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Releasing lock "refresh_cache-a3c01d4d-9c66-4441-95e8-87998782cc02" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.138026] env[62952]: DEBUG nova.compute.manager [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.138215] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.139164] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420dcbe0-9a3e-4c04-ab37-ae0e39c575e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.146891] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.147201] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-846e7355-b291-42e2-9b47-7c7d1fed75e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.154443] env[62952]: DEBUG oslo_vmware.api [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 957.154443] env[62952]: value = "task-1367305" [ 957.154443] env[62952]: _type = "Task" [ 957.154443] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.165212] env[62952]: DEBUG oslo_vmware.api [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.289764] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367304, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.292986] env[62952]: DEBUG nova.compute.manager [req-708d0e08-52b5-4f12-bdc6-fa97b866ea56 req-eaac2cee-089c-46a6-97a6-139c7931d818 service nova] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Received event network-vif-deleted-92eac1b0-516e-40aa-8e5b-5dc9299025fc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.344698] env[62952]: DEBUG oslo_concurrency.lockutils [req-b49a8453-bf06-4fea-ab16-d2eebb4d2dd2 req-b1d089fd-4f38-4709-9f52-020aa68ac1f4 service nova] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.363244] env[62952]: INFO nova.compute.manager [-] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Took 1.27 seconds to deallocate network for instance. [ 957.430654] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523fa0a6-0952-7fef-0ae5-7daa0d526700, 'name': SearchDatastore_Task, 'duration_secs': 0.011539} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.431703] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea60e76a-e410-493f-b109-c644dc0dd4f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.437653] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 957.437653] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52d9df58-0262-8bf9-2b51-8e07ec051dd0" [ 957.437653] env[62952]: _type = "Task" [ 957.437653] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.446037] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d9df58-0262-8bf9-2b51-8e07ec051dd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.525995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.928s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.529962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.397s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.529962] env[62952]: DEBUG nova.objects.instance [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 957.556661] env[62952]: INFO nova.scheduler.client.report [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Deleted allocations for instance 4983f354-77dc-4e84-969b-4c74441fa568 [ 957.632280] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 957.666870] env[62952]: DEBUG oslo_vmware.api [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367305, 'name': PowerOffVM_Task, 'duration_secs': 0.158012} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.667753] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.667977] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 957.668302] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b22edd23-e2e0-424b-a44b-6741f7c5084a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.691518] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 957.691774] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 957.691972] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Deleting the datastore file [datastore2] a3c01d4d-9c66-4441-95e8-87998782cc02 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.692294] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfb3b2f0-1d55-46a2-b22f-9266cafe7688 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.700263] env[62952]: DEBUG oslo_vmware.api [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for the task: (returnval){ [ 957.700263] env[62952]: value = "task-1367307" [ 957.700263] env[62952]: _type = "Task" [ 957.700263] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.709809] env[62952]: DEBUG oslo_vmware.api [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.792406] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367304, 'name': ReconfigVM_Task, 'duration_secs': 0.526508} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.792903] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Reconfigured VM instance instance-0000004b to attach disk [datastore2] ff189c9e-9e7c-4217-9c65-0f821393870a/ff189c9e-9e7c-4217-9c65-0f821393870a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.793688] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-782d7b13-14b1-4f08-8b38-f6071c554f19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.801756] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 957.801756] env[62952]: value = "task-1367308" [ 957.801756] env[62952]: _type = "Task" [ 957.801756] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.812097] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367308, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.872509] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.953183] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d9df58-0262-8bf9-2b51-8e07ec051dd0, 'name': SearchDatastore_Task, 'duration_secs': 0.015102} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.953747] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.954163] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 532a2bcd-5d6a-4fa9-abc9-a048cc915fda/532a2bcd-5d6a-4fa9-abc9-a048cc915fda.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.954555] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6d2dd4b-8981-4174-b9df-e71fd39bc30c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.963702] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 957.963702] env[62952]: value = "task-1367309" [ 957.963702] env[62952]: _type = "Task" [ 957.963702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.974693] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367309, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.064737] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0bbc9203-9520-4825-8e8d-603cdca9b33f tempest-ServerTagsTestJSON-2114128322 tempest-ServerTagsTestJSON-2114128322-project-member] Lock "4983f354-77dc-4e84-969b-4c74441fa568" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.301s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.097477] env[62952]: DEBUG nova.compute.manager [req-16083c91-822f-4e1d-86c8-2c61da403faa req-81840a08-d136-44f1-86f6-ff44be5bbbfe service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Received event network-vif-plugged-178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.098188] env[62952]: DEBUG oslo_concurrency.lockutils [req-16083c91-822f-4e1d-86c8-2c61da403faa req-81840a08-d136-44f1-86f6-ff44be5bbbfe service nova] Acquiring lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.098301] env[62952]: DEBUG oslo_concurrency.lockutils [req-16083c91-822f-4e1d-86c8-2c61da403faa req-81840a08-d136-44f1-86f6-ff44be5bbbfe service nova] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.098967] env[62952]: DEBUG oslo_concurrency.lockutils [req-16083c91-822f-4e1d-86c8-2c61da403faa req-81840a08-d136-44f1-86f6-ff44be5bbbfe service nova] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.099067] env[62952]: DEBUG nova.compute.manager [req-16083c91-822f-4e1d-86c8-2c61da403faa req-81840a08-d136-44f1-86f6-ff44be5bbbfe service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] No waiting events found dispatching network-vif-plugged-178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.099221] env[62952]: WARNING nova.compute.manager [req-16083c91-822f-4e1d-86c8-2c61da403faa req-81840a08-d136-44f1-86f6-ff44be5bbbfe service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Received unexpected event network-vif-plugged-178d3a5f-9120-4970-b88b-c7fe268836ca for instance with vm_state building and task_state spawning. [ 958.211491] env[62952]: DEBUG nova.network.neutron [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Successfully updated port: 178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 958.224401] env[62952]: DEBUG oslo_vmware.api [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Task: {'id': task-1367307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106033} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.225239] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.225539] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.225905] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.226282] env[62952]: INFO nova.compute.manager [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Took 1.09 seconds to destroy the instance on the hypervisor. [ 958.226752] env[62952]: DEBUG oslo.service.loopingcall [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.227513] env[62952]: DEBUG nova.compute.manager [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.227732] env[62952]: DEBUG nova.network.neutron [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.247436] env[62952]: DEBUG nova.network.neutron [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.314915] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367308, 'name': Rename_Task, 'duration_secs': 0.21379} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.315350] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.315654] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee115011-df3d-45e8-865f-b8b37eff274e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.325181] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 958.325181] env[62952]: value = "task-1367310" [ 958.325181] env[62952]: _type = "Task" [ 958.325181] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.335354] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.476118] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367309, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.540501] env[62952]: DEBUG oslo_concurrency.lockutils [None req-87cd51a2-ed56-47b4-8f9d-b79210f7bb50 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.541868] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.414s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.543043] env[62952]: INFO nova.compute.claims [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.715648] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "refresh_cache-8b2e4601-b981-48d1-84e6-21da66d5ea5f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.715648] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquired lock "refresh_cache-8b2e4601-b981-48d1-84e6-21da66d5ea5f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.715801] env[62952]: DEBUG nova.network.neutron [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.754573] env[62952]: DEBUG nova.network.neutron [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.836358] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367310, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.975669] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367309, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548398} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.976027] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 532a2bcd-5d6a-4fa9-abc9-a048cc915fda/532a2bcd-5d6a-4fa9-abc9-a048cc915fda.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 958.976299] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.976592] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d77842c4-5cd5-4b86-911a-93b88f6c3357 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.985417] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 958.985417] env[62952]: value = "task-1367311" [ 958.985417] env[62952]: _type = "Task" [ 958.985417] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.994940] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367311, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.257285] env[62952]: INFO nova.compute.manager [-] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Took 1.03 seconds to deallocate network for instance. [ 959.279834] env[62952]: DEBUG nova.network.neutron [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 959.338057] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367310, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.489981] env[62952]: DEBUG nova.network.neutron [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Updating instance_info_cache with network_info: [{"id": "178d3a5f-9120-4970-b88b-c7fe268836ca", "address": "fa:16:3e:7d:d2:c7", "network": {"id": "c9645289-4e0a-49dd-890f-5f3e6ce87a1f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-775701718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1396825e1044402b7d1c799657c1b94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap178d3a5f-91", "ovs_interfaceid": "178d3a5f-9120-4970-b88b-c7fe268836ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.497042] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367311, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096185} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.497842] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.498292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeddd036-d7df-4714-b6d5-073aab12cbba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.525272] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 532a2bcd-5d6a-4fa9-abc9-a048cc915fda/532a2bcd-5d6a-4fa9-abc9-a048cc915fda.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.525272] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e370d031-3fd9-4372-a6d7-531c89db5007 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.548333] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 959.548333] env[62952]: value = "task-1367312" [ 959.548333] env[62952]: _type = "Task" [ 959.548333] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.559637] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367312, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.768120] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.840299] env[62952]: DEBUG oslo_vmware.api [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367310, 'name': PowerOnVM_Task, 'duration_secs': 1.488978} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.844249] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.844611] env[62952]: INFO nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Took 9.07 seconds to spawn the instance on the hypervisor. [ 959.844879] env[62952]: DEBUG nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.846419] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7980a3fd-06e4-4551-9f28-b574c9e0774f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.899745] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f0b092-0454-4073-8a39-a1396707293f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.909704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e769c3df-00b7-4684-a317-ffac12405928 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.943345] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf50ee9-77e3-457a-bcc8-19a27a9fbcf7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.952161] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d429d5-89a5-4180-af2e-085dc3359336 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.969799] env[62952]: DEBUG nova.compute.provider_tree [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.993782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Releasing lock "refresh_cache-8b2e4601-b981-48d1-84e6-21da66d5ea5f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.993782] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Instance network_info: |[{"id": "178d3a5f-9120-4970-b88b-c7fe268836ca", "address": "fa:16:3e:7d:d2:c7", "network": {"id": "c9645289-4e0a-49dd-890f-5f3e6ce87a1f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-775701718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1396825e1044402b7d1c799657c1b94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap178d3a5f-91", "ovs_interfaceid": "178d3a5f-9120-4970-b88b-c7fe268836ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 960.060425] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367312, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.375362] env[62952]: INFO nova.compute.manager [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Took 27.42 seconds to build instance. [ 960.476039] env[62952]: DEBUG nova.scheduler.client.report [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.551188] env[62952]: DEBUG nova.compute.manager [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Received event network-changed-178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.551188] env[62952]: DEBUG nova.compute.manager [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Refreshing instance network info cache due to event network-changed-178d3a5f-9120-4970-b88b-c7fe268836ca. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.551188] env[62952]: DEBUG oslo_concurrency.lockutils [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] Acquiring lock "refresh_cache-8b2e4601-b981-48d1-84e6-21da66d5ea5f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.551188] env[62952]: DEBUG oslo_concurrency.lockutils [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] Acquired lock "refresh_cache-8b2e4601-b981-48d1-84e6-21da66d5ea5f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.551188] env[62952]: DEBUG nova.network.neutron [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Refreshing network info cache for port 178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.565454] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367312, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.877670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7c1bbb20-b685-462e-96ba-60ca1ec1b209 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.059s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.979958] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.980664] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.984630] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.849s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.984908] env[62952]: DEBUG nova.objects.instance [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lazy-loading 'resources' on Instance uuid 93951b09-9d7c-42da-83de-0db3980d0d5d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.068404] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367312, 'name': ReconfigVM_Task, 'duration_secs': 1.443496} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.068710] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 532a2bcd-5d6a-4fa9-abc9-a048cc915fda/532a2bcd-5d6a-4fa9-abc9-a048cc915fda.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.069413] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d9370d6-62d6-499d-b5dc-83fbc48e85e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.077517] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 961.077517] env[62952]: value = "task-1367313" [ 961.077517] env[62952]: _type = "Task" [ 961.077517] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.087933] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367313, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.362486] env[62952]: DEBUG nova.network.neutron [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Updated VIF entry in instance network info cache for port 178d3a5f-9120-4970-b88b-c7fe268836ca. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.362975] env[62952]: DEBUG nova.network.neutron [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Updating instance_info_cache with network_info: [{"id": "178d3a5f-9120-4970-b88b-c7fe268836ca", "address": "fa:16:3e:7d:d2:c7", "network": {"id": "c9645289-4e0a-49dd-890f-5f3e6ce87a1f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-775701718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1396825e1044402b7d1c799657c1b94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap178d3a5f-91", "ovs_interfaceid": "178d3a5f-9120-4970-b88b-c7fe268836ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.488910] env[62952]: DEBUG nova.compute.utils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.497926] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.498826] env[62952]: DEBUG nova.network.neutron [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.548643] env[62952]: DEBUG nova.policy [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17b50c91ccd347e4b45126a8f23c0278', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9657a113032417fb4b97716a772f8ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.593361] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367313, 'name': Rename_Task, 'duration_secs': 0.197751} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.596469] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.596963] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73275b2e-9a70-4e62-91f8-dcfc60b44244 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.604719] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 961.604719] env[62952]: value = "task-1367314" [ 961.604719] env[62952]: _type = "Task" [ 961.604719] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.617571] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367314, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.840207] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dd0337-be77-4bca-bfb3-4240d77f2f87 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.846560] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1ba766-1882-46b2-8377-cc5b70162368 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.888437] env[62952]: DEBUG oslo_concurrency.lockutils [req-5a6fe69c-0800-4f40-ab48-770095a49b66 req-51a5205d-7e4b-4031-b084-1ccd6410e2a7 service nova] Releasing lock "refresh_cache-8b2e4601-b981-48d1-84e6-21da66d5ea5f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.890754] env[62952]: DEBUG nova.network.neutron [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Successfully created port: 25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.893822] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d063b5a7-94ed-4757-a5b0-b7ffc2543c15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.903299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56b5163-a9f3-4eb5-8e8d-a75a416842b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.918172] env[62952]: DEBUG nova.compute.provider_tree [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.993507] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 962.098415] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.098415] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.098618] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.098973] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.099206] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.100263] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.100263] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.100263] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.100263] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.100263] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.100525] env[62952]: DEBUG nova.virt.hardware [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.102985] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba9eaf1-b8ce-4477-a226-5e842d68d0f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.116177] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f776ef-90e9-ddf6-aa4a-74f1e61ea052/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 962.117745] env[62952]: DEBUG nova.compute.manager [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-changed-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.117967] env[62952]: DEBUG nova.compute.manager [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing instance network info cache due to event network-changed-48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.118796] env[62952]: DEBUG oslo_concurrency.lockutils [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.119096] env[62952]: DEBUG oslo_concurrency.lockutils [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.119231] env[62952]: DEBUG nova.network.neutron [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.121827] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47758d98-2d1f-4839-ad8e-4c443d23ee35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.132141] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94156e6-51d8-4d43-86f0-7502ed667d81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.146102] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367314, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.146423] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f776ef-90e9-ddf6-aa4a-74f1e61ea052/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 962.146610] env[62952]: ERROR oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f776ef-90e9-ddf6-aa4a-74f1e61ea052/disk-0.vmdk due to incomplete transfer. [ 962.147231] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-89e38b62-a652-4e92-bfeb-8afdf2c4d99e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.157530] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:d2:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '178d3a5f-9120-4970-b88b-c7fe268836ca', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.165107] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Creating folder: Project (b1396825e1044402b7d1c799657c1b94). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 962.166424] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf5e2ac0-8b0a-4900-9ceb-87eebc76bf8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.172350] env[62952]: DEBUG oslo_vmware.rw_handles [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f776ef-90e9-ddf6-aa4a-74f1e61ea052/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 962.172560] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Uploaded image 1449561c-f27a-4b93-87d2-4669ae19d131 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 962.175650] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 962.176656] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-529f22fd-c317-4e45-81b7-950cdcc2784f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.180689] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Created folder: Project (b1396825e1044402b7d1c799657c1b94) in parent group-v290852. [ 962.180905] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Creating folder: Instances. Parent ref: group-v290930. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 962.181557] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78692c34-b163-477f-a230-e7ec6b9222a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.186423] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 962.186423] env[62952]: value = "task-1367316" [ 962.186423] env[62952]: _type = "Task" [ 962.186423] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.191753] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Created folder: Instances in parent group-v290930. [ 962.192074] env[62952]: DEBUG oslo.service.loopingcall [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.192292] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.192550] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40aed017-ab19-4ffa-9f86-2b740ba715d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.213011] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367316, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.218359] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.218359] env[62952]: value = "task-1367318" [ 962.218359] env[62952]: _type = "Task" [ 962.218359] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.226734] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367318, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.424022] env[62952]: DEBUG nova.scheduler.client.report [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.622291] env[62952]: DEBUG oslo_vmware.api [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367314, 'name': PowerOnVM_Task, 'duration_secs': 0.688801} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.622725] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.623185] env[62952]: INFO nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Took 9.41 seconds to spawn the instance on the hypervisor. [ 962.623649] env[62952]: DEBUG nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.625055] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0b115c-247f-4297-b39c-047d5c3cb4f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.697318] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367316, 'name': Destroy_Task, 'duration_secs': 0.481675} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.697565] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Destroyed the VM [ 962.697857] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 962.698179] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a58b055b-0c18-4a8f-8875-6d13af4417ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.705627] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 962.705627] env[62952]: value = "task-1367319" [ 962.705627] env[62952]: _type = "Task" [ 962.705627] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.714227] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367319, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.728346] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367318, 'name': CreateVM_Task, 'duration_secs': 0.384826} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.728885] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.729812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.729812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.730729] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.730729] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-331d9c5b-ccc3-4cf1-8f4e-5dc4a2f0091f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.736700] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 962.736700] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526310c4-50f7-00bc-2e52-99b563c7458f" [ 962.736700] env[62952]: _type = "Task" [ 962.736700] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.749456] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526310c4-50f7-00bc-2e52-99b563c7458f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.930440] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.946s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.932764] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.242s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.933042] env[62952]: DEBUG nova.objects.instance [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lazy-loading 'resources' on Instance uuid 7621d6a6-27aa-45d3-80fa-65e957519a1f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.957932] env[62952]: INFO nova.scheduler.client.report [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted allocations for instance 93951b09-9d7c-42da-83de-0db3980d0d5d [ 962.970265] env[62952]: DEBUG nova.network.neutron [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updated VIF entry in instance network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.970751] env[62952]: DEBUG nova.network.neutron [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.005643] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 963.029583] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.030089] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.030322] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.030827] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.031081] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.031308] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.031596] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.031801] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.032077] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.032287] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.032518] env[62952]: DEBUG nova.virt.hardware [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.033879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506de0ee-ef50-4818-a23f-2f54741c19fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.043690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a03993-055a-4393-941a-9b62d4abb7d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.153470] env[62952]: INFO nova.compute.manager [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Took 26.78 seconds to build instance. [ 963.154718] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.155011] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.155387] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.155603] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.155867] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.158862] env[62952]: INFO nova.compute.manager [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Terminating instance [ 963.161209] env[62952]: DEBUG nova.compute.manager [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.161407] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.162420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77aab1f-7be8-4ade-a59a-b688b2f30a3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.172058] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 963.172473] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8941ba4-8f72-4d52-81c1-dfe24d6e1deb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.216120] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.216632] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.224466] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367319, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.247398] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526310c4-50f7-00bc-2e52-99b563c7458f, 'name': SearchDatastore_Task, 'duration_secs': 0.012314} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.247610] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.247888] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.248298] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.248298] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.248529] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.248880] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f35ffe04-2b5a-47bf-a7cf-1f3ebd2e2ad5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.261422] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.261791] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.262769] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57149258-09d3-4f12-96ce-aa6765af81fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.270778] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 963.270778] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52403bcd-4616-d35a-a418-0011142b99ba" [ 963.270778] env[62952]: _type = "Task" [ 963.270778] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.281148] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52403bcd-4616-d35a-a418-0011142b99ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.449810] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 963.450257] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 963.450394] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleting the datastore file [datastore1] 94fd5287-9a8e-45e4-99e8-0bcc861f889c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.450701] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54f134eb-4a67-4f6e-8514-a81fb179c5e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.458527] env[62952]: DEBUG oslo_vmware.api [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 963.458527] env[62952]: value = "task-1367321" [ 963.458527] env[62952]: _type = "Task" [ 963.458527] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.469861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3073c827-6c55-419f-ace5-89cc5fef40ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "93951b09-9d7c-42da-83de-0db3980d0d5d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.812s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.478507] env[62952]: DEBUG oslo_concurrency.lockutils [req-af1def15-ad5a-4904-a907-e4598c3ee74b req-18437c29-bff7-4434-8a6a-1154996e43ed service nova] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.478926] env[62952]: DEBUG oslo_vmware.api [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.658283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a89598e9-42d4-4ae8-9610-9bc0668f9c96 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.273s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.716385] env[62952]: DEBUG oslo_vmware.api [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367319, 'name': RemoveSnapshot_Task, 'duration_secs': 0.593641} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.716748] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 963.716931] env[62952]: INFO nova.compute.manager [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Took 16.34 seconds to snapshot the instance on the hypervisor. [ 963.726520] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.748501] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bf6038-0130-40c5-b20c-161fcf1d64a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.760260] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879dbfa1-8a65-49ab-a93d-343a03d68214 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.805401] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcd2fa1-233a-4140-9bf4-5f9030a200d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.809125] env[62952]: DEBUG nova.compute.manager [req-6975a7a7-0154-444e-86cd-4bd4cb6859e7 req-8cdc9480-5b22-43db-a4b8-1f6791604b6a service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Received event network-vif-plugged-25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.809345] env[62952]: DEBUG oslo_concurrency.lockutils [req-6975a7a7-0154-444e-86cd-4bd4cb6859e7 req-8cdc9480-5b22-43db-a4b8-1f6791604b6a service nova] Acquiring lock "3bd9c356-42d6-450f-9069-b066952d9524-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.809553] env[62952]: DEBUG oslo_concurrency.lockutils [req-6975a7a7-0154-444e-86cd-4bd4cb6859e7 req-8cdc9480-5b22-43db-a4b8-1f6791604b6a service nova] Lock "3bd9c356-42d6-450f-9069-b066952d9524-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.809724] env[62952]: DEBUG oslo_concurrency.lockutils [req-6975a7a7-0154-444e-86cd-4bd4cb6859e7 req-8cdc9480-5b22-43db-a4b8-1f6791604b6a service nova] Lock "3bd9c356-42d6-450f-9069-b066952d9524-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.809896] env[62952]: DEBUG nova.compute.manager [req-6975a7a7-0154-444e-86cd-4bd4cb6859e7 req-8cdc9480-5b22-43db-a4b8-1f6791604b6a service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] No waiting events found dispatching network-vif-plugged-25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.810075] env[62952]: WARNING nova.compute.manager [req-6975a7a7-0154-444e-86cd-4bd4cb6859e7 req-8cdc9480-5b22-43db-a4b8-1f6791604b6a service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Received unexpected event network-vif-plugged-25128fb4-31cf-41ae-8d49-9fb9805d6044 for instance with vm_state building and task_state spawning. [ 963.820279] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1304e182-1918-4ce3-8035-f22f65536671 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.824124] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52403bcd-4616-d35a-a418-0011142b99ba, 'name': SearchDatastore_Task, 'duration_secs': 0.010224} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.825749] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a329459-3116-4edc-986b-efe53a6ebbd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.836711] env[62952]: DEBUG nova.compute.provider_tree [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.841796] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 963.841796] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526e25c5-3d63-99f6-4545-dfdf23b681d2" [ 963.841796] env[62952]: _type = "Task" [ 963.841796] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.854454] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526e25c5-3d63-99f6-4545-dfdf23b681d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.934858] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.938032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.979086] env[62952]: DEBUG oslo_vmware.api [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169064} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.979297] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.980144] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.980144] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.980528] env[62952]: INFO nova.compute.manager [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Took 0.82 seconds to destroy the instance on the hypervisor. [ 963.981289] env[62952]: DEBUG oslo.service.loopingcall [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.981289] env[62952]: DEBUG nova.compute.manager [-] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.981289] env[62952]: DEBUG nova.network.neutron [-] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.018834] env[62952]: DEBUG nova.network.neutron [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Successfully updated port: 25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 964.221100] env[62952]: DEBUG nova.compute.manager [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Instance disappeared during snapshot {{(pid=62952) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 964.244303] env[62952]: DEBUG nova.compute.manager [None req-37e62832-ab8c-4082-b326-d6bcfee0dba9 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image not found during clean up 1449561c-f27a-4b93-87d2-4669ae19d131 {{(pid=62952) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 964.258543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.339589] env[62952]: DEBUG nova.scheduler.client.report [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.365263] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526e25c5-3d63-99f6-4545-dfdf23b681d2, 'name': SearchDatastore_Task, 'duration_secs': 0.010908} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.366086] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.366493] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 8b2e4601-b981-48d1-84e6-21da66d5ea5f/8b2e4601-b981-48d1-84e6-21da66d5ea5f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 964.367414] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-388a614b-bec6-4c89-929f-c9100666e700 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.379059] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 964.379059] env[62952]: value = "task-1367322" [ 964.379059] env[62952]: _type = "Task" [ 964.379059] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.388927] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367322, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.440649] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.527450] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-3bd9c356-42d6-450f-9069-b066952d9524" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.527781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-3bd9c356-42d6-450f-9069-b066952d9524" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.531108] env[62952]: DEBUG nova.network.neutron [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.807214] env[62952]: DEBUG nova.compute.manager [req-6a44c261-63c1-41a7-8099-1121030ef87c req-ef35f3d8-0278-45be-85b8-360e32c9d91f service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Received event network-vif-deleted-b0c7b851-c731-4055-b489-db9ecae60444 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.807533] env[62952]: INFO nova.compute.manager [req-6a44c261-63c1-41a7-8099-1121030ef87c req-ef35f3d8-0278-45be-85b8-360e32c9d91f service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Neutron deleted interface b0c7b851-c731-4055-b489-db9ecae60444; detaching it from the instance and deleting it from the info cache [ 964.807785] env[62952]: DEBUG nova.network.neutron [req-6a44c261-63c1-41a7-8099-1121030ef87c req-ef35f3d8-0278-45be-85b8-360e32c9d91f service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.850517] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.855712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.168s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.855712] env[62952]: DEBUG nova.objects.instance [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'resources' on Instance uuid 34f5995f-692c-4f0d-8b15-0d388df7e34e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.890956] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367322, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.891698] env[62952]: INFO nova.scheduler.client.report [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted allocations for instance 7621d6a6-27aa-45d3-80fa-65e957519a1f [ 964.964796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.099349] env[62952]: DEBUG nova.network.neutron [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 965.245147] env[62952]: DEBUG nova.network.neutron [-] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.311964] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eee6fe36-44a9-4ab7-b474-be0659dc8f93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.324926] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cced36ac-fa57-4679-b86c-79fc8443be7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.364931] env[62952]: DEBUG nova.compute.manager [req-6a44c261-63c1-41a7-8099-1121030ef87c req-ef35f3d8-0278-45be-85b8-360e32c9d91f service nova] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Detach interface failed, port_id=b0c7b851-c731-4055-b489-db9ecae60444, reason: Instance 94fd5287-9a8e-45e4-99e8-0bcc861f889c could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 965.392319] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367322, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.392319] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 8b2e4601-b981-48d1-84e6-21da66d5ea5f/8b2e4601-b981-48d1-84e6-21da66d5ea5f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 965.392319] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.392710] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ab090c0-4925-45f4-877c-d3e31479992e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.402786] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 965.402786] env[62952]: value = "task-1367323" [ 965.402786] env[62952]: _type = "Task" [ 965.402786] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.407649] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1338059-8d31-49e9-a0af-a5161084cb13 tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "7621d6a6-27aa-45d3-80fa-65e957519a1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.898s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.420161] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.443641] env[62952]: DEBUG nova.network.neutron [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Updating instance_info_cache with network_info: [{"id": "25128fb4-31cf-41ae-8d49-9fb9805d6044", "address": "fa:16:3e:11:84:34", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25128fb4-31", "ovs_interfaceid": "25128fb4-31cf-41ae-8d49-9fb9805d6044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.679781] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecc5f7-757f-40ff-6ebd-450fc40574d3/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 965.681010] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992e2a96-e950-4e15-b0c0-cc349aa8af4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.692170] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecc5f7-757f-40ff-6ebd-450fc40574d3/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 965.692170] env[62952]: ERROR oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecc5f7-757f-40ff-6ebd-450fc40574d3/disk-0.vmdk due to incomplete transfer. [ 965.692170] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bf53ca0d-7f9c-480c-adb4-a510ae05ee58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.700477] env[62952]: DEBUG oslo_vmware.rw_handles [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ecc5f7-757f-40ff-6ebd-450fc40574d3/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 965.700477] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Uploaded image f9471984-046d-4570-b0fc-4791884b7f4f to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 965.701860] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 965.702145] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-77024eca-e7b7-45bd-898e-971c120865c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.708793] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 965.708793] env[62952]: value = "task-1367324" [ 965.708793] env[62952]: _type = "Task" [ 965.708793] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.719309] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367324, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.723609] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78b7bde-b9a1-4d84-9256-46a0d6f6879d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.735075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "da749732-72d8-4180-aedf-73a17fa9dea6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.735350] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "da749732-72d8-4180-aedf-73a17fa9dea6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.737462] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6585a1e3-f819-4db3-ba36-fb68b40ef672 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.777180] env[62952]: INFO nova.compute.manager [-] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Took 1.80 seconds to deallocate network for instance. [ 965.781105] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f03531-57fc-4ac4-ac5c-3a3ae8d336f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.793590] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385f2b32-9c63-432c-9914-cf8b5cae6e6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.812133] env[62952]: DEBUG nova.compute.provider_tree [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.834138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "1e02f16d-b86e-4533-928b-a672ce69788f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.834379] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.841025] env[62952]: DEBUG nova.compute.manager [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Received event network-changed-25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.841025] env[62952]: DEBUG nova.compute.manager [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Refreshing instance network info cache due to event network-changed-25128fb4-31cf-41ae-8d49-9fb9805d6044. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.841025] env[62952]: DEBUG oslo_concurrency.lockutils [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] Acquiring lock "refresh_cache-3bd9c356-42d6-450f-9069-b066952d9524" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.917215] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111929} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.917509] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.918392] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fbf3e2-68b3-4d0a-bb5e-0d00df12b8f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.943308] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 8b2e4601-b981-48d1-84e6-21da66d5ea5f/8b2e4601-b981-48d1-84e6-21da66d5ea5f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.943638] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67e43d9b-5b36-4fcd-b306-eb20928e2572 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.960521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-3bd9c356-42d6-450f-9069-b066952d9524" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.960948] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance network_info: |[{"id": "25128fb4-31cf-41ae-8d49-9fb9805d6044", "address": "fa:16:3e:11:84:34", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25128fb4-31", "ovs_interfaceid": "25128fb4-31cf-41ae-8d49-9fb9805d6044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.961190] env[62952]: DEBUG oslo_concurrency.lockutils [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] Acquired lock "refresh_cache-3bd9c356-42d6-450f-9069-b066952d9524" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.961339] env[62952]: DEBUG nova.network.neutron [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Refreshing network info cache for port 25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.962577] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:84:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25128fb4-31cf-41ae-8d49-9fb9805d6044', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.970428] env[62952]: DEBUG oslo.service.loopingcall [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.971353] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.972664] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f52bf7b6-010e-43e7-ba43-e74f6fe010aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.990270] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 965.990270] env[62952]: value = "task-1367325" [ 965.990270] env[62952]: _type = "Task" [ 965.990270] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.997357] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.997357] env[62952]: value = "task-1367326" [ 965.997357] env[62952]: _type = "Task" [ 965.997357] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.004621] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.009850] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367326, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.220854] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367324, 'name': Destroy_Task, 'duration_secs': 0.339611} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.221173] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Destroyed the VM [ 966.221440] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 966.221699] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-17a662e1-b750-417f-a8bb-561a7b0981a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.229301] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 966.229301] env[62952]: value = "task-1367327" [ 966.229301] env[62952]: _type = "Task" [ 966.229301] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.237936] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367327, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.243668] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 966.289361] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.316894] env[62952]: DEBUG nova.scheduler.client.report [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.337896] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 966.505097] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367325, 'name': ReconfigVM_Task, 'duration_secs': 0.318979} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.507075] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 8b2e4601-b981-48d1-84e6-21da66d5ea5f/8b2e4601-b981-48d1-84e6-21da66d5ea5f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.508097] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3c337ba-e626-44e0-97b6-66133004fd86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.517230] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367326, 'name': CreateVM_Task, 'duration_secs': 0.503652} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.517996] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.518860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.519196] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.524373] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.525455] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ceb6b7c-a767-4155-9ecc-bbfa1118b270 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.527257] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 966.527257] env[62952]: value = "task-1367328" [ 966.527257] env[62952]: _type = "Task" [ 966.527257] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.532141] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 966.532141] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5237db63-35be-302a-78a9-4ff9422bf21c" [ 966.532141] env[62952]: _type = "Task" [ 966.532141] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.540044] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367328, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.545760] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5237db63-35be-302a-78a9-4ff9422bf21c, 'name': SearchDatastore_Task, 'duration_secs': 0.011318} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.546036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.546289] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.546579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.546742] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.546926] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.547219] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4d8499e-8397-4dcb-be0a-fc11f29882e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.558245] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.558400] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.559150] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b59a5859-3026-468c-85e1-4555f1f91d8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.564718] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 966.564718] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52028427-8331-d0c7-deb8-f6d17920141a" [ 966.564718] env[62952]: _type = "Task" [ 966.564718] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.572582] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52028427-8331-d0c7-deb8-f6d17920141a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.701081] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "4c4845e6-79ca-4a2a-a636-842d6919e70f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.702035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.743496] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367327, 'name': RemoveSnapshot_Task, 'duration_secs': 0.343626} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.743890] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 966.744402] env[62952]: DEBUG nova.compute.manager [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.745296] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58739014-e69c-4e15-a838-545ea387ce72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.761960] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.764413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.777530] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.821996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.968s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.824423] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.596s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.828047] env[62952]: INFO nova.compute.claims [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.853173] env[62952]: INFO nova.scheduler.client.report [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocations for instance 34f5995f-692c-4f0d-8b15-0d388df7e34e [ 966.866116] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.912764] env[62952]: DEBUG nova.network.neutron [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Updated VIF entry in instance network info cache for port 25128fb4-31cf-41ae-8d49-9fb9805d6044. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.913207] env[62952]: DEBUG nova.network.neutron [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Updating instance_info_cache with network_info: [{"id": "25128fb4-31cf-41ae-8d49-9fb9805d6044", "address": "fa:16:3e:11:84:34", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25128fb4-31", "ovs_interfaceid": "25128fb4-31cf-41ae-8d49-9fb9805d6044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.038084] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367328, 'name': Rename_Task, 'duration_secs': 0.145616} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.038374] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.038632] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f8f2e96-6567-4c98-9121-8628996dbd7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.045803] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 967.045803] env[62952]: value = "task-1367329" [ 967.045803] env[62952]: _type = "Task" [ 967.045803] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.053538] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367329, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.074308] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52028427-8331-d0c7-deb8-f6d17920141a, 'name': SearchDatastore_Task, 'duration_secs': 0.01001} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.075118] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20d86385-9451-48a7-bdfe-bdee641f7990 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.080451] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 967.080451] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52fbd757-eb7d-7135-db60-e2c708b6e8ba" [ 967.080451] env[62952]: _type = "Task" [ 967.080451] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.087988] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fbd757-eb7d-7135-db60-e2c708b6e8ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.205520] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.269630] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.278073] env[62952]: INFO nova.compute.manager [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Shelve offloading [ 967.279697] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.279968] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e028c665-c3ef-4b28-bb71-3d4cc527b5e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.287823] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 967.287823] env[62952]: value = "task-1367330" [ 967.287823] env[62952]: _type = "Task" [ 967.287823] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.296393] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.364187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-add4dc72-96dc-4244-a27c-25ff47be7cb8 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "34f5995f-692c-4f0d-8b15-0d388df7e34e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.830s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.415995] env[62952]: DEBUG oslo_concurrency.lockutils [req-5e72428e-e15a-4bcc-8a2b-09cbb9729c97 req-a08188a0-73fc-4a3b-8f85-d8093e57426c service nova] Releasing lock "refresh_cache-3bd9c356-42d6-450f-9069-b066952d9524" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.557027] env[62952]: DEBUG oslo_vmware.api [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367329, 'name': PowerOnVM_Task, 'duration_secs': 0.478853} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.557027] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.557245] env[62952]: INFO nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Took 9.92 seconds to spawn the instance on the hypervisor. [ 967.557438] env[62952]: DEBUG nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.558244] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca8cb66-6349-4807-8bb5-93d565bd29a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.590275] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fbd757-eb7d-7135-db60-e2c708b6e8ba, 'name': SearchDatastore_Task, 'duration_secs': 0.019027} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.590548] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.590820] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.591096] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0618784b-9c42-4bc1-b72e-20d216c97295 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.597751] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 967.597751] env[62952]: value = "task-1367331" [ 967.597751] env[62952]: _type = "Task" [ 967.597751] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.608013] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.725403] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.794181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.803450] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 967.803450] env[62952]: DEBUG nova.compute.manager [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.804410] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5e7995-1910-456f-aa66-a92ac8bf1394 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.814059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.814059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.814059] env[62952]: DEBUG nova.network.neutron [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.930018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.930273] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.078922] env[62952]: INFO nova.compute.manager [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Took 28.64 seconds to build instance. [ 968.113276] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367331, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.166420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f45315-2fa8-4675-8acf-6fddf6e32282 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.177098] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec4f34b-04a0-49e1-b87c-bc56a0ec6aa2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.211082] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac631746-92e1-4a48-864d-14a5145d5ed0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.221317] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e24e4d-4d7e-48e5-a9cd-2ded011dcf45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.239463] env[62952]: DEBUG nova.compute.provider_tree [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.532529] env[62952]: DEBUG nova.network.neutron [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updating instance_info_cache with network_info: [{"id": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "address": "fa:16:3e:f0:ce:a3", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034e9fd2-98", "ovs_interfaceid": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.581458] env[62952]: DEBUG oslo_concurrency.lockutils [None req-13002e03-c10c-4b8f-ba80-a6b11f5a1865 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.713s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.611265] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367331, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691582} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.611545] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.611779] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.612045] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-134320cc-f6b3-4750-8222-e9aba0f16038 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.621040] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 968.621040] env[62952]: value = "task-1367332" [ 968.621040] env[62952]: _type = "Task" [ 968.621040] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.629550] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367332, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.746235] env[62952]: DEBUG nova.scheduler.client.report [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.035485] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.083870] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 969.131760] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367332, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069717} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.132057] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.132862] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad04a12-4c00-4286-818f-b47b688f3029 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.161644] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.161981] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0521cc66-a535-4f6f-843c-f97dd11b576e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.183533] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 969.183533] env[62952]: value = "task-1367333" [ 969.183533] env[62952]: _type = "Task" [ 969.183533] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.193502] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.252271] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.252796] env[62952]: DEBUG nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.255565] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.710s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.257199] env[62952]: INFO nova.compute.claims [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.305379] env[62952]: DEBUG nova.compute.manager [req-70d46b34-0d17-42f1-ae16-eea4352bfe6a req-010f6de2-cac3-4a75-ba52-6c4c4a2995bd service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Received event network-vif-unplugged-034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.305379] env[62952]: DEBUG oslo_concurrency.lockutils [req-70d46b34-0d17-42f1-ae16-eea4352bfe6a req-010f6de2-cac3-4a75-ba52-6c4c4a2995bd service nova] Acquiring lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.305379] env[62952]: DEBUG oslo_concurrency.lockutils [req-70d46b34-0d17-42f1-ae16-eea4352bfe6a req-010f6de2-cac3-4a75-ba52-6c4c4a2995bd service nova] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.305379] env[62952]: DEBUG oslo_concurrency.lockutils [req-70d46b34-0d17-42f1-ae16-eea4352bfe6a req-010f6de2-cac3-4a75-ba52-6c4c4a2995bd service nova] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.305598] env[62952]: DEBUG nova.compute.manager [req-70d46b34-0d17-42f1-ae16-eea4352bfe6a req-010f6de2-cac3-4a75-ba52-6c4c4a2995bd service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] No waiting events found dispatching network-vif-unplugged-034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.306305] env[62952]: WARNING nova.compute.manager [req-70d46b34-0d17-42f1-ae16-eea4352bfe6a req-010f6de2-cac3-4a75-ba52-6c4c4a2995bd service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Received unexpected event network-vif-unplugged-034e9fd2-98bf-4d19-a01e-627fcfc75b89 for instance with vm_state shelved and task_state shelving_offloading. [ 969.327544] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.328514] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0d8146-4260-4cb3-9f1e-f42d0af7e34f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.337458] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.337702] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-762e8710-26f7-481c-ada9-f9ed8f954659 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.576652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.578618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.578618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.578618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.578618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.580134] env[62952]: INFO nova.compute.manager [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Terminating instance [ 969.582176] env[62952]: DEBUG nova.compute.manager [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.582392] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.583626] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7246633f-efe6-43e2-ab32-3c32d975f3c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.592178] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.592454] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.592658] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleting the datastore file [datastore2] 5ace8d6f-49ab-4486-bee1-04cd24f7f430 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.592939] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-abadcb6f-097e-485f-b696-567156ec3cc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.600117] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.601024] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c826b3c-2133-4ff0-963f-6632e2bd40cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.607305] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 969.607305] env[62952]: value = "task-1367335" [ 969.607305] env[62952]: _type = "Task" [ 969.607305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.609209] env[62952]: DEBUG oslo_vmware.api [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 969.609209] env[62952]: value = "task-1367336" [ 969.609209] env[62952]: _type = "Task" [ 969.609209] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.623230] env[62952]: DEBUG oslo_vmware.api [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.626883] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367335, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.636181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.695481] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.766669] env[62952]: DEBUG nova.compute.utils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 969.769425] env[62952]: DEBUG nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 970.122313] env[62952]: DEBUG oslo_vmware.api [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367335, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183761} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.125212] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.125413] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.125598] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.127524] env[62952]: DEBUG oslo_vmware.api [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367336, 'name': PowerOffVM_Task, 'duration_secs': 0.191218} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.128044] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.128236] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.128649] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4b95593-219c-493e-8ba7-ff1ba15ea137 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.160447] env[62952]: INFO nova.scheduler.client.report [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocations for instance 5ace8d6f-49ab-4486-bee1-04cd24f7f430 [ 970.196947] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367333, 'name': ReconfigVM_Task, 'duration_secs': 0.883394} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.197299] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.198007] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8f92dea-25ec-49b4-8934-34056318b3ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.205778] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 970.205778] env[62952]: value = "task-1367338" [ 970.205778] env[62952]: _type = "Task" [ 970.205778] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.215085] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367338, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.233615] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.233893] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.234134] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Deleting the datastore file [datastore2] 8b2e4601-b981-48d1-84e6-21da66d5ea5f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.234575] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5479d6bc-526e-48e6-9dbc-c227123660bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.245977] env[62952]: DEBUG oslo_vmware.api [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for the task: (returnval){ [ 970.245977] env[62952]: value = "task-1367339" [ 970.245977] env[62952]: _type = "Task" [ 970.245977] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.256113] env[62952]: DEBUG oslo_vmware.api [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.270492] env[62952]: DEBUG nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.593378] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18a3bae-0604-49fe-bbc3-7d397e5cc906 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.601908] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9331203-0c60-4601-a93f-8e776b77c295 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.634639] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b55d7db-8aa4-421c-8794-59eceb68b62f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.643334] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5098ced-8133-4a06-aa93-ad0d8f6cd6f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.657721] env[62952]: DEBUG nova.compute.provider_tree [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.666850] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.716920] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367338, 'name': Rename_Task, 'duration_secs': 0.182808} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.717266] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 970.717520] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f70360c-cbcd-4fee-b776-dc2250bd157e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.725138] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 970.725138] env[62952]: value = "task-1367340" [ 970.725138] env[62952]: _type = "Task" [ 970.725138] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.733600] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367340, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.757017] env[62952]: DEBUG oslo_vmware.api [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Task: {'id': task-1367339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145624} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.757336] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.757476] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.757695] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.757940] env[62952]: INFO nova.compute.manager [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 970.758161] env[62952]: DEBUG oslo.service.loopingcall [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.758382] env[62952]: DEBUG nova.compute.manager [-] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.758478] env[62952]: DEBUG nova.network.neutron [-] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 971.161319] env[62952]: DEBUG nova.scheduler.client.report [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.235834] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367340, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.283056] env[62952]: DEBUG nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.305842] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.305842] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.305842] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.306063] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.306309] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.306446] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.306667] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.306835] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.307096] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.307226] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.307430] env[62952]: DEBUG nova.virt.hardware [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.308340] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc70b4a5-ea4b-42ec-bab0-625d33166fc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.319009] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4091320-4ec2-4752-a529-320137d1488e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.334061] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.338886] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Creating folder: Project (6a4834c416ee4dc58e47e0caac58561f). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.339230] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-355122ca-b80e-4fa6-b718-9aecd0e82a1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.351342] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Created folder: Project (6a4834c416ee4dc58e47e0caac58561f) in parent group-v290852. [ 971.351568] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Creating folder: Instances. Parent ref: group-v290934. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.351820] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1b95b29-0a28-471f-b077-28fbffb1d04b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.362851] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Created folder: Instances in parent group-v290934. [ 971.363146] env[62952]: DEBUG oslo.service.loopingcall [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.363322] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.363532] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-abe2ffa4-8f9f-417e-bfa2-ef4d1bb3f8b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.381713] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.381713] env[62952]: value = "task-1367343" [ 971.381713] env[62952]: _type = "Task" [ 971.381713] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.389743] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367343, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.441784] env[62952]: DEBUG nova.compute.manager [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Received event network-changed-034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.442155] env[62952]: DEBUG nova.compute.manager [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Refreshing instance network info cache due to event network-changed-034e9fd2-98bf-4d19-a01e-627fcfc75b89. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 971.442543] env[62952]: DEBUG oslo_concurrency.lockutils [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] Acquiring lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.442782] env[62952]: DEBUG oslo_concurrency.lockutils [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] Acquired lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.442939] env[62952]: DEBUG nova.network.neutron [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Refreshing network info cache for port 034e9fd2-98bf-4d19-a01e-627fcfc75b89 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.629836] env[62952]: DEBUG nova.network.neutron [-] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.666657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.667237] env[62952]: DEBUG nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 971.670178] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.349s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.738954] env[62952]: DEBUG oslo_vmware.api [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367340, 'name': PowerOnVM_Task, 'duration_secs': 0.564581} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.739130] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.739472] env[62952]: INFO nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Took 8.73 seconds to spawn the instance on the hypervisor. [ 971.739472] env[62952]: DEBUG nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.740244] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a25bf5a-ffe5-4a63-841c-8cc7b946e2d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.891303] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367343, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.905109] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.132688] env[62952]: INFO nova.compute.manager [-] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Took 1.37 seconds to deallocate network for instance. [ 972.173967] env[62952]: DEBUG nova.compute.utils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.182680] env[62952]: DEBUG nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Not allocating networking since 'none' was specified. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 972.219506] env[62952]: DEBUG nova.network.neutron [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updated VIF entry in instance network info cache for port 034e9fd2-98bf-4d19-a01e-627fcfc75b89. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.219844] env[62952]: DEBUG nova.network.neutron [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updating instance_info_cache with network_info: [{"id": "034e9fd2-98bf-4d19-a01e-627fcfc75b89", "address": "fa:16:3e:f0:ce:a3", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": null, "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap034e9fd2-98", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.258128] env[62952]: INFO nova.compute.manager [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Took 28.16 seconds to build instance. [ 972.392832] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367343, 'name': CreateVM_Task, 'duration_secs': 0.573772} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.392998] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 972.393459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.393628] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.394023] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 972.394275] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9debcefb-38ee-410a-adb0-b54778be80de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.398997] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 972.398997] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52dd6ac2-99c4-0d5e-7133-c2151cba741e" [ 972.398997] env[62952]: _type = "Task" [ 972.398997] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.407673] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52dd6ac2-99c4-0d5e-7133-c2151cba741e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.643013] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.684036] env[62952]: DEBUG nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 972.723089] env[62952]: DEBUG oslo_concurrency.lockutils [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] Releasing lock "refresh_cache-5ace8d6f-49ab-4486-bee1-04cd24f7f430" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.723378] env[62952]: DEBUG nova.compute.manager [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Received event network-vif-deleted-178d3a5f-9120-4970-b88b-c7fe268836ca {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.723609] env[62952]: INFO nova.compute.manager [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Neutron deleted interface 178d3a5f-9120-4970-b88b-c7fe268836ca; detaching it from the instance and deleting it from the info cache [ 972.723938] env[62952]: DEBUG nova.network.neutron [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.735320] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a63ec2ac-4484-4360-962b-105a5338c1b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.735643] env[62952]: WARNING nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 802fd770-38aa-4417-b25e-fd83510bceef is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 972.738024] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738024] env[62952]: WARNING nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 94fd5287-9a8e-45e4-99e8-0bcc861f889c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 972.738024] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738024] env[62952]: WARNING nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a3c01d4d-9c66-4441-95e8-87998782cc02 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 972.738024] env[62952]: WARNING nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance a88447f6-6a5a-490d-bf44-11c9237fa5cb is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 972.738473] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738473] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ff189c9e-9e7c-4217-9c65-0f821393870a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738473] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 532a2bcd-5d6a-4fa9-abc9-a048cc915fda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738473] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 8b2e4601-b981-48d1-84e6-21da66d5ea5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738636] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 3bd9c356-42d6-450f-9069-b066952d9524 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738636] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 82401449-347e-4917-a307-d5d7bc048f7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.738636] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 97995e38-b27e-478a-8553-eb1c844bb0a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 972.759534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d0557a78-b0bc-4008-90b9-edab3cd6b48d tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "3bd9c356-42d6-450f-9069-b066952d9524" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.677s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.910417] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52dd6ac2-99c4-0d5e-7133-c2151cba741e, 'name': SearchDatastore_Task, 'duration_secs': 0.013619} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.910725] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.910956] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.911249] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.911411] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.911595] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.911858] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ab3e4d5-f299-4ea1-8d4a-e69aedd6b84c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.921042] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.921244] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.921961] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c4c667d-975f-44b3-813e-04dfb54c6631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.927429] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 972.927429] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a1c2b4-00e4-d009-f288-b4454557232d" [ 972.927429] env[62952]: _type = "Task" [ 972.927429] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.936418] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a1c2b4-00e4-d009-f288-b4454557232d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.227636] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ea2d20f-07e2-4180-9865-39b05ed101f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.238753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7ef65a-89b9-4945-9d6e-4b09dea55a35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.249394] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 303c9fe7-0d89-4a1a-adcc-6f787486df61 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 973.269339] env[62952]: DEBUG nova.compute.manager [req-58cb897f-d88c-49a2-88b7-6f0d8e62c456 req-9354902d-c25f-4338-b4bc-af8ee3b737f1 service nova] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Detach interface failed, port_id=178d3a5f-9120-4970-b88b-c7fe268836ca, reason: Instance 8b2e4601-b981-48d1-84e6-21da66d5ea5f could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 973.439792] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a1c2b4-00e4-d009-f288-b4454557232d, 'name': SearchDatastore_Task, 'duration_secs': 0.009372} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.440601] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87251871-9721-4f64-8658-dc48b3d1894f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.445808] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 973.445808] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5271f5c1-6743-bfaf-e3d0-55a1d7baf756" [ 973.445808] env[62952]: _type = "Task" [ 973.445808] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.453256] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5271f5c1-6743-bfaf-e3d0-55a1d7baf756, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.578742] env[62952]: INFO nova.compute.manager [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Rebuilding instance [ 973.616392] env[62952]: DEBUG nova.compute.manager [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.617302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663b9fde-609c-43a3-9a8d-2d2b56cc828a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.697958] env[62952]: DEBUG nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 973.718652] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.718910] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.719086] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.719272] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.719423] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.719573] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.719779] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.719938] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.720121] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.720289] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.720462] env[62952]: DEBUG nova.virt.hardware [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.721585] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bae155-40c9-46b6-9e25-02227521c5c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.729906] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d9f056-abfd-4d5a-8e1a-64fb2bb0e7d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.744353] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.749782] env[62952]: DEBUG oslo.service.loopingcall [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.750017] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.750237] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0c0cca4-7069-408f-ad13-42082e99617f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.762644] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 244f7b73-fda4-4477-b04d-c4f5cedd2d98 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 973.771024] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.771024] env[62952]: value = "task-1367344" [ 973.771024] env[62952]: _type = "Task" [ 973.771024] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.779699] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367344, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.957439] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5271f5c1-6743-bfaf-e3d0-55a1d7baf756, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.957835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.958338] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 82401449-347e-4917-a307-d5d7bc048f7c/82401449-347e-4917-a307-d5d7bc048f7c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.958726] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0934e7ae-ead1-4d45-9f51-69f0349a1a50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.968075] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 973.968075] env[62952]: value = "task-1367345" [ 973.968075] env[62952]: _type = "Task" [ 973.968075] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.976461] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.131370] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.131370] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-553209d9-e332-423b-92c6-ed42b9baf34b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.138826] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 974.138826] env[62952]: value = "task-1367346" [ 974.138826] env[62952]: _type = "Task" [ 974.138826] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.149517] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.265741] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance da749732-72d8-4180-aedf-73a17fa9dea6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 974.284098] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367344, 'name': CreateVM_Task, 'duration_secs': 0.271732} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.284098] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.284291] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.284657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.285334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.285955] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-513c8722-2b5c-4754-8a29-dc5626758838 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.293969] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 974.293969] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a38825-6396-79bb-b5a0-8b57a2efda29" [ 974.293969] env[62952]: _type = "Task" [ 974.293969] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.304805] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a38825-6396-79bb-b5a0-8b57a2efda29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.477348] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447052} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.477618] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 82401449-347e-4917-a307-d5d7bc048f7c/82401449-347e-4917-a307-d5d7bc048f7c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.477835] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.478102] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5190e1f7-bd2a-4030-863b-848a28932bab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.484956] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 974.484956] env[62952]: value = "task-1367347" [ 974.484956] env[62952]: _type = "Task" [ 974.484956] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.494129] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.649752] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367346, 'name': PowerOffVM_Task, 'duration_secs': 0.241451} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.650057] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.650299] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.651086] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5942ca45-a26b-4cc3-8554-e92f43150f30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.659180] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.659450] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1692bad9-c936-4578-be86-b2201a694d88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.750468] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.750726] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.750913] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.751212] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16fa40b0-8a69-4eb0-bfcc-871efbada1a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.758740] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 974.758740] env[62952]: value = "task-1367349" [ 974.758740] env[62952]: _type = "Task" [ 974.758740] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.767241] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.769129] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 1e02f16d-b86e-4533-928b-a672ce69788f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 974.804116] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a38825-6396-79bb-b5a0-8b57a2efda29, 'name': SearchDatastore_Task, 'duration_secs': 0.064582} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.804431] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.804663] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.804899] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.805063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.805252] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.805517] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-202b6680-ee4c-4696-a747-fd13cd3c0b92 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.814384] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.814595] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.815310] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce062b14-3ec0-435c-b22b-7f5e8561b743 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.821356] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 974.821356] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52db3a61-747c-c929-2f02-9f81f8c00c11" [ 974.821356] env[62952]: _type = "Task" [ 974.821356] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.829703] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52db3a61-747c-c929-2f02-9f81f8c00c11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.995758] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367347, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069588} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.996201] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.996977] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55615d7f-ee02-45f3-8103-88742b1ac636 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.016844] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 82401449-347e-4917-a307-d5d7bc048f7c/82401449-347e-4917-a307-d5d7bc048f7c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.017144] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e0436f9-2212-44eb-9100-5bfc3e245e8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.037350] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 975.037350] env[62952]: value = "task-1367350" [ 975.037350] env[62952]: _type = "Task" [ 975.037350] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.046164] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.269129] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148397} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.269498] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.269635] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.269816] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.272715] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 4c4845e6-79ca-4a2a-a636-842d6919e70f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 975.333631] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52db3a61-747c-c929-2f02-9f81f8c00c11, 'name': SearchDatastore_Task, 'duration_secs': 0.010464} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.334376] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a4d4690-8027-4c4d-8445-907633bb8c49 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.340245] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 975.340245] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52aee97a-e1ea-87f8-08d6-3113515d1df2" [ 975.340245] env[62952]: _type = "Task" [ 975.340245] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.350409] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52aee97a-e1ea-87f8-08d6-3113515d1df2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.548168] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367350, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.779343] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ba2dfb85-0ffa-4883-abf1-482441c5afb8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 975.851397] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52aee97a-e1ea-87f8-08d6-3113515d1df2, 'name': SearchDatastore_Task, 'duration_secs': 0.009935} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.851684] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.852023] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.852298] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9165e1b3-5652-44c0-846f-3bf5333f614c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.859727] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 975.859727] env[62952]: value = "task-1367351" [ 975.859727] env[62952]: _type = "Task" [ 975.859727] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.867707] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.047727] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367350, 'name': ReconfigVM_Task, 'duration_secs': 0.614392} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.048087] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 82401449-347e-4917-a307-d5d7bc048f7c/82401449-347e-4917-a307-d5d7bc048f7c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.048718] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90f9a6e3-27cd-4f37-a02c-8bed6a6773a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.055591] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 976.055591] env[62952]: value = "task-1367352" [ 976.055591] env[62952]: _type = "Task" [ 976.055591] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.064285] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367352, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.284513] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 75d640ed-c41a-4761-8867-191d8b3e1f79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 976.284849] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 976.285038] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 976.307770] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.307901] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.308022] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.308212] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.308639] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.308639] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.308791] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.308905] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.309072] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.309206] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.309383] env[62952]: DEBUG nova.virt.hardware [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.310282] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d68df9-8159-46f1-b6c9-47ba54d4d5cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.321515] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7323813-1856-415a-8d5c-958f41d81774 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.338202] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:84:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25128fb4-31cf-41ae-8d49-9fb9805d6044', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.345456] env[62952]: DEBUG oslo.service.loopingcall [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.347796] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 976.348191] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32377082-fca3-495f-9929-dd8136e9f75c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.372166] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.375904] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.375904] env[62952]: value = "task-1367353" [ 976.375904] env[62952]: _type = "Task" [ 976.375904] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.383726] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367353, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.567771] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367352, 'name': Rename_Task, 'duration_secs': 0.200297} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.569081] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.569835] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92861954-e993-4d0f-aa98-0d01cc329f4c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.572275] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c766b0b-1da3-45cd-8d3d-603bca4dd52a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.324429] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6540a78-1b73-4b06-9f7f-8f946babba67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.327466] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 977.327466] env[62952]: value = "task-1367354" [ 977.327466] env[62952]: _type = "Task" [ 977.327466] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.359767] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.359984] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367353, 'name': CreateVM_Task, 'duration_secs': 0.478387} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.361397] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 977.362855] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb093518-025d-4025-a401-91ecf2c9e86f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.365089] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.365265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.365577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.367273] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5f33265-70ea-4b7a-8c12-f096940c8d25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.378461] env[62952]: DEBUG oslo_vmware.api [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367354, 'name': PowerOnVM_Task, 'duration_secs': 0.487576} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.378777] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 977.378777] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c1e7ab-07bd-f87f-cbcc-02b69449b7eb" [ 977.378777] env[62952]: _type = "Task" [ 977.378777] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.379035] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 977.379257] env[62952]: INFO nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Took 6.10 seconds to spawn the instance on the hypervisor. [ 977.379459] env[62952]: DEBUG nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.380856] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4750751d-e31b-432c-9111-e0c0b8fc62c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.385157] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c595af67-993a-414d-9366-6462f854f6e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.398016] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c1e7ab-07bd-f87f-cbcc-02b69449b7eb, 'name': SearchDatastore_Task, 'duration_secs': 0.012057} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.408273] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.409017] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.409017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.409017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.409159] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.409527] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.410832] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb8a3e5b-7138-4c83-9c81-ff1c944b44ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.420249] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.420560] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 977.422176] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e955be26-4de2-4e18-b156-1a260a8ecdc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.427069] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 977.427069] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c01fa0-111a-2c1c-1af1-006438c7688b" [ 977.427069] env[62952]: _type = "Task" [ 977.427069] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.437779] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c01fa0-111a-2c1c-1af1-006438c7688b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.825562] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.915612] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.920419] env[62952]: INFO nova.compute.manager [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Took 27.71 seconds to build instance. [ 977.938434] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c01fa0-111a-2c1c-1af1-006438c7688b, 'name': SearchDatastore_Task, 'duration_secs': 0.01141} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.939781] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01d8ad7f-51cb-4e7f-a25c-b41021055b19 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.945889] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 977.945889] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]525030b8-b830-c80a-191e-e09f50d1ec2f" [ 977.945889] env[62952]: _type = "Task" [ 977.945889] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.954221] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525030b8-b830-c80a-191e-e09f50d1ec2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.327868] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367351, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.421893] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 978.422279] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.752s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.422835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-69f1e2cf-9550-479a-a4f2-945bec52d432 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "82401449-347e-4917-a307-d5d7bc048f7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.226s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.423203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.798s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.423548] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.426191] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.101s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.426483] env[62952]: DEBUG nova.objects.instance [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 978.450920] env[62952]: INFO nova.scheduler.client.report [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Deleted allocations for instance a88447f6-6a5a-490d-bf44-11c9237fa5cb [ 978.461010] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525030b8-b830-c80a-191e-e09f50d1ec2f, 'name': SearchDatastore_Task, 'duration_secs': 0.011954} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.461605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.461931] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.462526] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d067fb5-8336-40ca-abc0-d8b2d4194920 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.471655] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 978.471655] env[62952]: value = "task-1367355" [ 978.471655] env[62952]: _type = "Task" [ 978.471655] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.482490] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367355, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.829303] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367351, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.670424} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.829605] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.829825] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.830103] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18334846-9b58-461a-bcec-44bc0b169e95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.839877] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 978.839877] env[62952]: value = "task-1367356" [ 978.839877] env[62952]: _type = "Task" [ 978.839877] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.849526] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.965521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b08e0e2f-5f92-4dac-873e-6e0d9f04f830 tempest-InstanceActionsNegativeTestJSON-963076843 tempest-InstanceActionsNegativeTestJSON-963076843-project-member] Lock "a88447f6-6a5a-490d-bf44-11c9237fa5cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.245s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.985727] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367355, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.351497] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070857} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.351769] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.352591] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c773c7-8e57-482c-a60f-da2e6b8a7625 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.372893] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.373172] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cffa707-00e5-41f6-92f1-5385c0b1921e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.394231] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 979.394231] env[62952]: value = "task-1367357" [ 979.394231] env[62952]: _type = "Task" [ 979.394231] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.402199] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367357, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.436593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fe72745-bc31-4c09-813b-909b97ca8ac0 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.437769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.565s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.437974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.440070] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.672s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.440284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.441974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.184s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.443801] env[62952]: INFO nova.compute.claims [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.460864] env[62952]: INFO nova.scheduler.client.report [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleted allocations for instance 802fd770-38aa-4417-b25e-fd83510bceef [ 979.466896] env[62952]: INFO nova.scheduler.client.report [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Deleted allocations for instance a3c01d4d-9c66-4441-95e8-87998782cc02 [ 979.486492] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367355, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637214} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.486958] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.487215] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.487593] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b67c8c6-6140-4d48-8466-010c16e25d9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.498023] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 979.498023] env[62952]: value = "task-1367358" [ 979.498023] env[62952]: _type = "Task" [ 979.498023] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.509422] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367358, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.905097] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367357, 'name': ReconfigVM_Task, 'duration_secs': 0.332648} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.905417] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.906295] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b08174cf-789a-482b-8acd-c81e0f97cf63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.914884] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 979.914884] env[62952]: value = "task-1367359" [ 979.914884] env[62952]: _type = "Task" [ 979.914884] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.924286] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367359, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.976978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7244efb9-f6fb-4a43-b0cf-82ac325a3b3d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "802fd770-38aa-4417-b25e-fd83510bceef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.162s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.980101] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16ec032b-2984-4af6-922e-5bba1d470c38 tempest-ServerShowV257Test-1294082725 tempest-ServerShowV257Test-1294082725-project-member] Lock "a3c01d4d-9c66-4441-95e8-87998782cc02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.965s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.010190] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367358, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071684} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.010606] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.011535] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fd5f73-1231-45c5-bdad-f5beca071ecd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.036774] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.037578] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01e26a80-d1fa-440c-b55f-1076b18cf8df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.061478] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 980.061478] env[62952]: value = "task-1367360" [ 980.061478] env[62952]: _type = "Task" [ 980.061478] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.073511] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367360, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.427179] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367359, 'name': Rename_Task, 'duration_secs': 0.163313} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.427475] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.427763] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f9cc93a-5ba3-4a72-85f4-493ead1b9632 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.435029] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 980.435029] env[62952]: value = "task-1367361" [ 980.435029] env[62952]: _type = "Task" [ 980.435029] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.443845] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367361, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.577941] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.771849] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07b1d46-5c4c-4660-a0f3-65c139725f82 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.780774] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdeff00-9035-44b0-bb50-f2b07f29715f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.813871] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d087787-0dc8-43b6-8dbe-9290136a5383 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.825962] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de545d1b-c66f-43cc-afa2-9bb91024619d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.843313] env[62952]: DEBUG nova.compute.provider_tree [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.947814] env[62952]: DEBUG oslo_vmware.api [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367361, 'name': PowerOnVM_Task, 'duration_secs': 0.466307} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.947977] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.948207] env[62952]: INFO nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Took 7.25 seconds to spawn the instance on the hypervisor. [ 980.948432] env[62952]: DEBUG nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.950079] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b719aca-8136-4dc1-9ee1-b4be552d8fe3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.076369] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367360, 'name': ReconfigVM_Task, 'duration_secs': 0.740285} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.076599] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524/3bd9c356-42d6-450f-9069-b066952d9524.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.077233] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c34c1e7-bdc0-4436-8b8a-81f2acb24bc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.087101] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 981.087101] env[62952]: value = "task-1367362" [ 981.087101] env[62952]: _type = "Task" [ 981.087101] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.095644] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367362, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.168362] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "a63ec2ac-4484-4360-962b-105a5338c1b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.168884] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.169134] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "a63ec2ac-4484-4360-962b-105a5338c1b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.169329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.169506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.173837] env[62952]: INFO nova.compute.manager [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Terminating instance [ 981.175736] env[62952]: DEBUG nova.compute.manager [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.175941] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.176808] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafe7475-b299-4f8c-8b99-5ce43f24e0ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.185494] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.185731] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2c07689-a261-4f35-bb9d-89dbc6961412 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.193399] env[62952]: DEBUG oslo_vmware.api [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 981.193399] env[62952]: value = "task-1367363" [ 981.193399] env[62952]: _type = "Task" [ 981.193399] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.202194] env[62952]: DEBUG oslo_vmware.api [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.349899] env[62952]: DEBUG nova.scheduler.client.report [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.478070] env[62952]: INFO nova.compute.manager [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Took 30.95 seconds to build instance. [ 981.598492] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367362, 'name': Rename_Task, 'duration_secs': 0.23539} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.598787] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.599058] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95552828-d479-4a91-a7a6-f6c4e44825e4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.608340] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 981.608340] env[62952]: value = "task-1367364" [ 981.608340] env[62952]: _type = "Task" [ 981.608340] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.617860] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367364, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.705179] env[62952]: DEBUG oslo_vmware.api [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367363, 'name': PowerOffVM_Task, 'duration_secs': 0.188726} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.705548] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.705646] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.705899] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d776a6ed-4fab-4c9d-bfa5-65bc65677d0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.855789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.856620] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.860418] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.896s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.866453] env[62952]: INFO nova.compute.claims [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.980920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91e9e182-1a2c-45a1-b92b-8ae4c2e2b319 tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "97995e38-b27e-478a-8553-eb1c844bb0a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.459s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.981337] env[62952]: INFO nova.compute.manager [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Rebuilding instance [ 982.025573] env[62952]: DEBUG nova.compute.manager [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.026505] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d857c46-4bd4-4ba0-96c3-265cff416e97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.118367] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367364, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.372433] env[62952]: DEBUG nova.compute.utils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.375834] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 982.376037] env[62952]: DEBUG nova.network.neutron [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 982.414887] env[62952]: DEBUG nova.policy [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb541b4abd9426ea96edcd066c69be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d1f373267464f14ae2c833151821973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.538433] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.538763] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b116be21-7c1a-434c-b333-2b2101eb9cf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.550729] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 982.550729] env[62952]: value = "task-1367366" [ 982.550729] env[62952]: _type = "Task" [ 982.550729] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.563484] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367366, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.618910] env[62952]: DEBUG oslo_vmware.api [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367364, 'name': PowerOnVM_Task, 'duration_secs': 0.895653} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.619247] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.619864] env[62952]: DEBUG nova.compute.manager [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.620310] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295fe9b4-5a34-4fbf-9927-c37501d5d2bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.716116] env[62952]: DEBUG nova.network.neutron [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Successfully created port: 3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.880064] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 983.062516] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367366, 'name': PowerOffVM_Task, 'duration_secs': 0.122029} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.062691] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.062873] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.064354] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06943315-af8e-41da-abea-9fbb633eb86d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.070653] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.072842] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a899c4b5-9559-40f0-8cf6-eb8758fa045e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.100253] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.100501] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.100750] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleting the datastore file [datastore1] 97995e38-b27e-478a-8553-eb1c844bb0a6 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.101049] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79ec40f7-3e8c-4b6d-bbc6-60e8a62b4afb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.107305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5714b11-770c-447f-b15f-3df5f4cbf27e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.110921] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 983.110921] env[62952]: value = "task-1367368" [ 983.110921] env[62952]: _type = "Task" [ 983.110921] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.116968] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a0cdc0-7c02-470d-8446-52fb54e46b25 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.123485] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.169088] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cf818b-0587-499c-b047-d676daa05588 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.175494] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.181498] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82f61d6-80aa-4a71-ba2c-58074d38f475 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.196050] env[62952]: DEBUG nova.compute.provider_tree [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.621354] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091144} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.621715] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.621889] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.622069] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.699977] env[62952]: DEBUG nova.scheduler.client.report [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.843276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.843522] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.893148] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.922023] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.922023] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.922023] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.922267] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.922267] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.922267] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.922267] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.922412] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.922611] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.922859] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.923019] env[62952]: DEBUG nova.virt.hardware [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.924246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5665fde1-ab94-442f-b322-9ee45eb89c54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.932908] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b4ec4a-7220-4a59-a461-9ccf5dab1cad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.066147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "3bd9c356-42d6-450f-9069-b066952d9524" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.066147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "3bd9c356-42d6-450f-9069-b066952d9524" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.066147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "3bd9c356-42d6-450f-9069-b066952d9524-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.066147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "3bd9c356-42d6-450f-9069-b066952d9524-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.066388] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "3bd9c356-42d6-450f-9069-b066952d9524-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.067019] env[62952]: INFO nova.compute.manager [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Terminating instance [ 984.073156] env[62952]: DEBUG nova.compute.manager [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.076017] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.076017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a31657-7f79-4810-8d86-333c90a71549 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.084140] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.084140] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3f004a1-90e1-4ba8-a10c-622ceb89805c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.091251] env[62952]: DEBUG oslo_vmware.api [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 984.091251] env[62952]: value = "task-1367369" [ 984.091251] env[62952]: _type = "Task" [ 984.091251] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.100400] env[62952]: DEBUG oslo_vmware.api [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.205021] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.205021] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 984.208159] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.919s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.208442] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.210435] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.433s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.212208] env[62952]: INFO nova.compute.claims [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.215425] env[62952]: DEBUG nova.compute.manager [req-98cd44ef-8a10-4003-a3ca-beff1ba42159 req-cc268700-a10c-4e57-b7c1-f2c2de8ee633 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-vif-plugged-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.215702] env[62952]: DEBUG oslo_concurrency.lockutils [req-98cd44ef-8a10-4003-a3ca-beff1ba42159 req-cc268700-a10c-4e57-b7c1-f2c2de8ee633 service nova] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.216072] env[62952]: DEBUG oslo_concurrency.lockutils [req-98cd44ef-8a10-4003-a3ca-beff1ba42159 req-cc268700-a10c-4e57-b7c1-f2c2de8ee633 service nova] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.216445] env[62952]: DEBUG oslo_concurrency.lockutils [req-98cd44ef-8a10-4003-a3ca-beff1ba42159 req-cc268700-a10c-4e57-b7c1-f2c2de8ee633 service nova] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.216996] env[62952]: DEBUG nova.compute.manager [req-98cd44ef-8a10-4003-a3ca-beff1ba42159 req-cc268700-a10c-4e57-b7c1-f2c2de8ee633 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] No waiting events found dispatching network-vif-plugged-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.217295] env[62952]: WARNING nova.compute.manager [req-98cd44ef-8a10-4003-a3ca-beff1ba42159 req-cc268700-a10c-4e57-b7c1-f2c2de8ee633 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received unexpected event network-vif-plugged-3a39f7b2-d402-4ead-8172-34ae1a9c4387 for instance with vm_state building and task_state spawning. [ 984.242069] env[62952]: INFO nova.scheduler.client.report [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted allocations for instance 94fd5287-9a8e-45e4-99e8-0bcc861f889c [ 984.308257] env[62952]: DEBUG nova.network.neutron [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Successfully updated port: 3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.346907] env[62952]: DEBUG nova.compute.utils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.491965] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.492220] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.601863] env[62952]: DEBUG oslo_vmware.api [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367369, 'name': PowerOffVM_Task, 'duration_secs': 0.190534} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.602833] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.602833] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.602833] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-137edb3b-a881-47b9-9fbf-7ad7f42b5bf7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.652896] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.652896] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.653309] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.653309] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.653413] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.653573] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.653783] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.653949] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.654137] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.654311] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.654489] env[62952]: DEBUG nova.virt.hardware [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.655400] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474b2e5d-9ab3-4089-9a93-a44ffb6ba59f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.663793] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f21d4f-d855-47e2-8788-45d7067f1d5f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.678913] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance VIF info [] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.684572] env[62952]: DEBUG oslo.service.loopingcall [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.685854] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.686146] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.686374] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.686564] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore2] 3bd9c356-42d6-450f-9069-b066952d9524 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.686795] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-421674da-7d72-417d-9fe1-dc8e218f5dd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.698687] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af02ef48-ae4f-43b9-8902-a4d7a4f07a1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.708819] env[62952]: DEBUG oslo_vmware.api [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 984.708819] env[62952]: value = "task-1367371" [ 984.708819] env[62952]: _type = "Task" [ 984.708819] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.710045] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.710045] env[62952]: value = "task-1367372" [ 984.710045] env[62952]: _type = "Task" [ 984.710045] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.723889] env[62952]: DEBUG nova.compute.utils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.725568] env[62952]: DEBUG oslo_vmware.api [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.729302] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.729522] env[62952]: DEBUG nova.network.neutron [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.731508] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367372, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.753538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f0d2e95c-8c7b-4e2d-8d9d-9cc2fc70b23c tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "94fd5287-9a8e-45e4-99e8-0bcc861f889c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.598s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.797865] env[62952]: DEBUG nova.policy [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55ffb58a3b3a4ad29699d48ecd6c0700', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d564dc3c75e430dbaf1f90a9c90c18b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.811063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.811265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.811661] env[62952]: DEBUG nova.network.neutron [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.849447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.995229] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 985.100067] env[62952]: DEBUG nova.network.neutron [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Successfully created port: fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.224698] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367372, 'name': CreateVM_Task, 'duration_secs': 0.320808} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.227500] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.230187] env[62952]: DEBUG oslo_vmware.api [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192011} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.230730] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.230994] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.231324] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.231484] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.231658] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.231838] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.232023] env[62952]: INFO nova.compute.manager [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Took 1.16 seconds to destroy the instance on the hypervisor. [ 985.232257] env[62952]: DEBUG oslo.service.loopingcall [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.232490] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca429b41-6cd4-430a-8789-17a50ddf17fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.234132] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 985.236613] env[62952]: DEBUG nova.compute.manager [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.236728] env[62952]: DEBUG nova.network.neutron [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 985.242472] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 985.242472] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b06e52-3d56-e3eb-7366-28a96b24ec92" [ 985.242472] env[62952]: _type = "Task" [ 985.242472] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.250747] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b06e52-3d56-e3eb-7366-28a96b24ec92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.356752] env[62952]: DEBUG nova.network.neutron [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 985.377855] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.378920] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.378920] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleting the datastore file [datastore1] a63ec2ac-4484-4360-962b-105a5338c1b8 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.379063] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b23a451e-f68f-4a8d-86cb-c2fee7ecbb15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.392123] env[62952]: DEBUG oslo_vmware.api [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for the task: (returnval){ [ 985.392123] env[62952]: value = "task-1367373" [ 985.392123] env[62952]: _type = "Task" [ 985.392123] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.401923] env[62952]: DEBUG oslo_vmware.api [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367373, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.494540] env[62952]: DEBUG nova.compute.manager [req-a2cedff6-2851-4277-828a-3cd555898a33 req-4f0552ea-b5db-4b75-a038-7335c7bdccd0 service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Received event network-vif-deleted-25128fb4-31cf-41ae-8d49-9fb9805d6044 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.494804] env[62952]: INFO nova.compute.manager [req-a2cedff6-2851-4277-828a-3cd555898a33 req-4f0552ea-b5db-4b75-a038-7335c7bdccd0 service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Neutron deleted interface 25128fb4-31cf-41ae-8d49-9fb9805d6044; detaching it from the instance and deleting it from the info cache [ 985.495029] env[62952]: DEBUG nova.network.neutron [req-a2cedff6-2851-4277-828a-3cd555898a33 req-4f0552ea-b5db-4b75-a038-7335c7bdccd0 service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.516793] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.519203] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2c6650-e2a5-4ee6-b6cd-41e5c3e8d53f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.527988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d55e0f-06b6-469b-a7f0-bf10a759767b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.561197] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60207bbe-8e33-4bab-bdee-827765aed8d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.568852] env[62952]: DEBUG nova.network.neutron [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.571857] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d82f9b1-9e4f-435f-a0b0-9429e643461b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.587431] env[62952]: DEBUG nova.compute.provider_tree [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.754085] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b06e52-3d56-e3eb-7366-28a96b24ec92, 'name': SearchDatastore_Task, 'duration_secs': 0.013894} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.754507] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.754800] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.755071] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.755236] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.755460] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.755709] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50f2914e-0fbb-4584-a648-8490a2797790 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.767870] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.768089] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.769644] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cc38be1-556e-41a4-aad8-f11511895d07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.779928] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 985.779928] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279dce4-5b50-fce2-be10-a899899174f6" [ 985.779928] env[62952]: _type = "Task" [ 985.779928] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.791450] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5279dce4-5b50-fce2-be10-a899899174f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.904291] env[62952]: DEBUG oslo_vmware.api [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Task: {'id': task-1367373, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165595} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.904600] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.904787] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.904924] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.905118] env[62952]: INFO nova.compute.manager [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Took 4.73 seconds to destroy the instance on the hypervisor. [ 985.905368] env[62952]: DEBUG oslo.service.loopingcall [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.905562] env[62952]: DEBUG nova.compute.manager [-] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.905653] env[62952]: DEBUG nova.network.neutron [-] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 985.946096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.946332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.946564] env[62952]: INFO nova.compute.manager [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Attaching volume 5b018b5e-81cf-4820-85bd-d3015cbda494 to /dev/sdb [ 985.966418] env[62952]: DEBUG nova.network.neutron [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.981819] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0533e4e5-7f21-4821-aec0-32616708b76b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.991677] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce0707d-36a3-4aa3-ad6f-302ca36740d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.000972] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adcb3924-86d6-432f-a7f6-cc9f6e6c1883 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.006087] env[62952]: DEBUG nova.virt.block_device [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating existing volume attachment record: c09185ee-fae6-404f-b3b5-18e9967ebfb9 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 986.015675] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb5222b-61a3-4148-a4b3-50e92b8cb8f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.043838] env[62952]: DEBUG nova.compute.manager [req-a2cedff6-2851-4277-828a-3cd555898a33 req-4f0552ea-b5db-4b75-a038-7335c7bdccd0 service nova] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Detach interface failed, port_id=25128fb4-31cf-41ae-8d49-9fb9805d6044, reason: Instance 3bd9c356-42d6-450f-9069-b066952d9524 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 986.071189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.071732] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Instance network_info: |[{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.072185] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:34:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a39f7b2-d402-4ead-8172-34ae1a9c4387', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.079660] env[62952]: DEBUG oslo.service.loopingcall [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.079900] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.080138] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba3b95db-d2e2-4fd5-8c30-b741b7ccaedd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.094793] env[62952]: DEBUG nova.scheduler.client.report [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.106346] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.106346] env[62952]: value = "task-1367374" [ 986.106346] env[62952]: _type = "Task" [ 986.106346] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.116634] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367374, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.240399] env[62952]: DEBUG nova.compute.manager [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.240655] env[62952]: DEBUG nova.compute.manager [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing instance network info cache due to event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 986.240919] env[62952]: DEBUG oslo_concurrency.lockutils [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.241528] env[62952]: DEBUG oslo_concurrency.lockutils [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.241528] env[62952]: DEBUG nova.network.neutron [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 986.243606] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 986.273233] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.273552] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.273552] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.273621] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.273797] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.273914] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.274133] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.274299] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.274881] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.274881] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.274881] env[62952]: DEBUG nova.virt.hardware [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.280122] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7af771-4386-4842-b4ba-2e05d52b647d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.296114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a3cf1c-ad69-434d-b748-827eb0f39c9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.300907] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5279dce4-5b50-fce2-be10-a899899174f6, 'name': SearchDatastore_Task, 'duration_secs': 0.012271} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.302128] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5c650ab-369e-4cad-afb7-f30720e25e8e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.317016] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 986.317016] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]529af836-20a6-d70d-a28b-5507992db85f" [ 986.317016] env[62952]: _type = "Task" [ 986.317016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.326396] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529af836-20a6-d70d-a28b-5507992db85f, 'name': SearchDatastore_Task, 'duration_secs': 0.010498} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.326830] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.327200] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.327494] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b3639fc-4ecd-4d98-980c-e1c64005a91b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.335841] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 986.335841] env[62952]: value = "task-1367377" [ 986.335841] env[62952]: _type = "Task" [ 986.335841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.345072] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.469126] env[62952]: INFO nova.compute.manager [-] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Took 1.23 seconds to deallocate network for instance. [ 986.600221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.600819] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.603851] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.738s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.606087] env[62952]: INFO nova.compute.claims [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.619827] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367374, 'name': CreateVM_Task, 'duration_secs': 0.366526} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.620058] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.620892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.621305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.622833] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.622833] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6997195-8523-4027-9252-7812af414d6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.630035] env[62952]: DEBUG nova.network.neutron [-] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.631287] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 986.631287] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52521857-1355-c258-6fea-ed5d58d76b1b" [ 986.631287] env[62952]: _type = "Task" [ 986.631287] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.642649] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52521857-1355-c258-6fea-ed5d58d76b1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.790867] env[62952]: DEBUG nova.network.neutron [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Successfully updated port: fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.847115] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479079} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.847758] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.847989] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.848270] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1ac9f7c-c4fd-41d1-81bb-86c615283d86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.857702] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 986.857702] env[62952]: value = "task-1367379" [ 986.857702] env[62952]: _type = "Task" [ 986.857702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.866932] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367379, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.978859] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.019515] env[62952]: DEBUG nova.network.neutron [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updated VIF entry in instance network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.020057] env[62952]: DEBUG nova.network.neutron [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.111023] env[62952]: DEBUG nova.compute.utils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.115066] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.115066] env[62952]: DEBUG nova.network.neutron [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.135794] env[62952]: INFO nova.compute.manager [-] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Took 1.23 seconds to deallocate network for instance. [ 987.145107] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52521857-1355-c258-6fea-ed5d58d76b1b, 'name': SearchDatastore_Task, 'duration_secs': 0.056667} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.145568] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.145796] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.146044] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.146199] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.146381] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.146635] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-906c82a6-2997-48ba-ad1c-83e252dc0516 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.153172] env[62952]: DEBUG nova.policy [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.167822] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.168082] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.169343] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6480e97-3c7c-4628-9978-98749978d45f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.175823] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 987.175823] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52645dcb-26cc-25b0-f7a5-7dd631ac38ae" [ 987.175823] env[62952]: _type = "Task" [ 987.175823] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.184014] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52645dcb-26cc-25b0-f7a5-7dd631ac38ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.293824] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "refresh_cache-244f7b73-fda4-4477-b04d-c4f5cedd2d98" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.294051] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "refresh_cache-244f7b73-fda4-4477-b04d-c4f5cedd2d98" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.294142] env[62952]: DEBUG nova.network.neutron [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.368557] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367379, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125921} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.369525] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.370149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a70d89-1670-4fa6-b54f-a2bd649dae8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.397202] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.397672] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64f892a9-3b74-4947-b29c-626e872b40e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.421749] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 987.421749] env[62952]: value = "task-1367380" [ 987.421749] env[62952]: _type = "Task" [ 987.421749] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.432206] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.449705] env[62952]: DEBUG nova.network.neutron [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Successfully created port: 3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.523366] env[62952]: DEBUG oslo_concurrency.lockutils [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.523499] env[62952]: DEBUG nova.compute.manager [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Received event network-vif-deleted-cea71c6a-30a0-4f56-bc79-6b103c4b59c3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.523693] env[62952]: INFO nova.compute.manager [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Neutron deleted interface cea71c6a-30a0-4f56-bc79-6b103c4b59c3; detaching it from the instance and deleting it from the info cache [ 987.523945] env[62952]: DEBUG nova.network.neutron [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.615742] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.646748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.689740] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52645dcb-26cc-25b0-f7a5-7dd631ac38ae, 'name': SearchDatastore_Task, 'duration_secs': 0.019716} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.696810] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c1f3885-85ff-4411-beec-77c124137da4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.699403] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 987.699403] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52093e4d-2347-cbb9-b1be-729b6cac21d6" [ 987.699403] env[62952]: _type = "Task" [ 987.699403] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.711667] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52093e4d-2347-cbb9-b1be-729b6cac21d6, 'name': SearchDatastore_Task, 'duration_secs': 0.010243} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.711992] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.712211] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 303c9fe7-0d89-4a1a-adcc-6f787486df61/303c9fe7-0d89-4a1a-adcc-6f787486df61.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.712466] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eda7194-45c5-4a2f-9999-bfc50bfd8ae2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.719365] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 987.719365] env[62952]: value = "task-1367381" [ 987.719365] env[62952]: _type = "Task" [ 987.719365] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.729867] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.834064] env[62952]: DEBUG nova.network.neutron [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.868650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0edfecd-3e6d-4159-9a0d-b437c5cc8fba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.877150] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031dc241-519f-4473-898e-875367bde735 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.912050] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3e4eed-e52e-4e8b-b4c9-0c34611185d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.923026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e754fa5-a590-4bf2-80d6-192da9884ddb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.942202] env[62952]: DEBUG nova.compute.provider_tree [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.946120] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367380, 'name': ReconfigVM_Task, 'duration_secs': 0.285184} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.946674] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 97995e38-b27e-478a-8553-eb1c844bb0a6/97995e38-b27e-478a-8553-eb1c844bb0a6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.947391] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab5dfd48-44ab-42d8-b3ed-61febf33aa09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.957735] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 987.957735] env[62952]: value = "task-1367382" [ 987.957735] env[62952]: _type = "Task" [ 987.957735] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.974235] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367382, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.028776] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c23f752-b5b4-485b-aa64-843848a24e1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.040085] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3695e466-0ab2-4285-b88f-7630c72b841b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.071554] env[62952]: DEBUG nova.compute.manager [req-8d278afb-4a14-43c7-b015-40c3867641d2 req-6afc03c0-fd92-40c4-83db-2d0e1a95c66d service nova] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Detach interface failed, port_id=cea71c6a-30a0-4f56-bc79-6b103c4b59c3, reason: Instance a63ec2ac-4484-4360-962b-105a5338c1b8 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 988.090797] env[62952]: DEBUG nova.network.neutron [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Updating instance_info_cache with network_info: [{"id": "fe087c06-3240-4db7-9b5d-ec61d2d983c0", "address": "fa:16:3e:87:c4:34", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe087c06-32", "ovs_interfaceid": "fe087c06-3240-4db7-9b5d-ec61d2d983c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.233884] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506791} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.234202] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 303c9fe7-0d89-4a1a-adcc-6f787486df61/303c9fe7-0d89-4a1a-adcc-6f787486df61.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.234426] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.234768] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8bae940-0a09-4373-b6eb-5a7fc465dc37 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.242865] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 988.242865] env[62952]: value = "task-1367383" [ 988.242865] env[62952]: _type = "Task" [ 988.242865] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.251932] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.268835] env[62952]: DEBUG nova.compute.manager [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Received event network-vif-plugged-fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.269082] env[62952]: DEBUG oslo_concurrency.lockutils [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] Acquiring lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.269293] env[62952]: DEBUG oslo_concurrency.lockutils [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.269478] env[62952]: DEBUG oslo_concurrency.lockutils [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.269643] env[62952]: DEBUG nova.compute.manager [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] No waiting events found dispatching network-vif-plugged-fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.269813] env[62952]: WARNING nova.compute.manager [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Received unexpected event network-vif-plugged-fe087c06-3240-4db7-9b5d-ec61d2d983c0 for instance with vm_state building and task_state spawning. [ 988.270011] env[62952]: DEBUG nova.compute.manager [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Received event network-changed-fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.270152] env[62952]: DEBUG nova.compute.manager [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Refreshing instance network info cache due to event network-changed-fe087c06-3240-4db7-9b5d-ec61d2d983c0. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.270326] env[62952]: DEBUG oslo_concurrency.lockutils [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] Acquiring lock "refresh_cache-244f7b73-fda4-4477-b04d-c4f5cedd2d98" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.448045] env[62952]: DEBUG nova.scheduler.client.report [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.470242] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367382, 'name': Rename_Task, 'duration_secs': 0.291165} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.471156] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.471421] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-935a89d4-0ec0-4523-823f-93c8d5c703fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.480579] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 988.480579] env[62952]: value = "task-1367385" [ 988.480579] env[62952]: _type = "Task" [ 988.480579] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.496969] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367385, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.593194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "refresh_cache-244f7b73-fda4-4477-b04d-c4f5cedd2d98" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.593754] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Instance network_info: |[{"id": "fe087c06-3240-4db7-9b5d-ec61d2d983c0", "address": "fa:16:3e:87:c4:34", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe087c06-32", "ovs_interfaceid": "fe087c06-3240-4db7-9b5d-ec61d2d983c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 988.594116] env[62952]: DEBUG oslo_concurrency.lockutils [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] Acquired lock "refresh_cache-244f7b73-fda4-4477-b04d-c4f5cedd2d98" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.594317] env[62952]: DEBUG nova.network.neutron [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Refreshing network info cache for port fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.595619] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:c4:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92fe29b3-0907-453d-aabb-5559c4bd7c0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe087c06-3240-4db7-9b5d-ec61d2d983c0', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.603463] env[62952]: DEBUG oslo.service.loopingcall [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.606352] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.606859] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d882c4d0-9eee-441d-a57c-d4d83c435898 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.627838] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.631295] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.631295] env[62952]: value = "task-1367386" [ 988.631295] env[62952]: _type = "Task" [ 988.631295] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.640657] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367386, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.654941] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.655216] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.655384] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.655572] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.655722] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.655875] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.656122] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.656325] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.656509] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.656676] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.656879] env[62952]: DEBUG nova.virt.hardware [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.658072] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6e0d3b-5d59-4cf0-bf3e-870327128efc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.667704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afb572b-9ecb-4e06-b1eb-9713ab33e684 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.753537] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072692} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.756285] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.757981] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6259afb7-b857-4f00-bc9f-7138d2630994 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.782167] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 303c9fe7-0d89-4a1a-adcc-6f787486df61/303c9fe7-0d89-4a1a-adcc-6f787486df61.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.782626] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cce7f9c-0f7d-4101-93f2-003f04621d51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.804728] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 988.804728] env[62952]: value = "task-1367387" [ 988.804728] env[62952]: _type = "Task" [ 988.804728] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.816309] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367387, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.941901] env[62952]: DEBUG nova.network.neutron [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Updated VIF entry in instance network info cache for port fe087c06-3240-4db7-9b5d-ec61d2d983c0. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.942420] env[62952]: DEBUG nova.network.neutron [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Updating instance_info_cache with network_info: [{"id": "fe087c06-3240-4db7-9b5d-ec61d2d983c0", "address": "fa:16:3e:87:c4:34", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe087c06-32", "ovs_interfaceid": "fe087c06-3240-4db7-9b5d-ec61d2d983c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.949757] env[62952]: DEBUG nova.compute.manager [req-45b5b557-1d5e-4d45-bbba-b1b1b3ba0fe1 req-c578e8e1-65a3-48bb-aa8c-d4f20340ffcc service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Received event network-vif-plugged-3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.949988] env[62952]: DEBUG oslo_concurrency.lockutils [req-45b5b557-1d5e-4d45-bbba-b1b1b3ba0fe1 req-c578e8e1-65a3-48bb-aa8c-d4f20340ffcc service nova] Acquiring lock "da749732-72d8-4180-aedf-73a17fa9dea6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.950214] env[62952]: DEBUG oslo_concurrency.lockutils [req-45b5b557-1d5e-4d45-bbba-b1b1b3ba0fe1 req-c578e8e1-65a3-48bb-aa8c-d4f20340ffcc service nova] Lock "da749732-72d8-4180-aedf-73a17fa9dea6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.950381] env[62952]: DEBUG oslo_concurrency.lockutils [req-45b5b557-1d5e-4d45-bbba-b1b1b3ba0fe1 req-c578e8e1-65a3-48bb-aa8c-d4f20340ffcc service nova] Lock "da749732-72d8-4180-aedf-73a17fa9dea6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.950545] env[62952]: DEBUG nova.compute.manager [req-45b5b557-1d5e-4d45-bbba-b1b1b3ba0fe1 req-c578e8e1-65a3-48bb-aa8c-d4f20340ffcc service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] No waiting events found dispatching network-vif-plugged-3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.950713] env[62952]: WARNING nova.compute.manager [req-45b5b557-1d5e-4d45-bbba-b1b1b3ba0fe1 req-c578e8e1-65a3-48bb-aa8c-d4f20340ffcc service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Received unexpected event network-vif-plugged-3582454d-1ad8-4562-8f55-42a3f083d328 for instance with vm_state building and task_state spawning. [ 988.953153] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.953694] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.956730] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.231s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.958227] env[62952]: INFO nova.compute.claims [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.994197] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367385, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.142450] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367386, 'name': CreateVM_Task, 'duration_secs': 0.432163} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.142668] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.143410] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.143625] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.143944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.144216] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-204dc27b-c2b5-468d-8fea-be894e215185 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.149051] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 989.149051] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]520a94c5-868b-6d9c-f579-957c6870c9cf" [ 989.149051] env[62952]: _type = "Task" [ 989.149051] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.157185] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520a94c5-868b-6d9c-f579-957c6870c9cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.318585] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367387, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.445602] env[62952]: DEBUG oslo_concurrency.lockutils [req-41c01bdf-d1e3-40eb-b293-65d69496f341 req-33777b7a-567d-4f15-97e9-b9fa26e3b170 service nova] Releasing lock "refresh_cache-244f7b73-fda4-4477-b04d-c4f5cedd2d98" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.465884] env[62952]: DEBUG nova.compute.utils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.467896] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.468184] env[62952]: DEBUG nova.network.neutron [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 989.492164] env[62952]: DEBUG oslo_vmware.api [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367385, 'name': PowerOnVM_Task, 'duration_secs': 0.513999} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.492452] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.492669] env[62952]: DEBUG nova.compute.manager [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.493486] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9238dc27-bf49-4a69-b6c9-393dda3142cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.521165] env[62952]: DEBUG nova.policy [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '913eacf6a94845b189e0bd5df408138a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8c2a296e8fe4805b9bb761f91a407f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.563445] env[62952]: DEBUG nova.network.neutron [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Successfully updated port: 3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.660213] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520a94c5-868b-6d9c-f579-957c6870c9cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009503} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.660536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.660794] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.661052] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.661216] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.661400] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.661658] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2217640c-75b5-4052-9dab-2b39b0a284a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.670069] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.670256] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.670924] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb57cde2-8807-41e6-9a9e-b9bf94617dea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.675925] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 989.675925] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b238c2-21f2-ce2d-bd1e-f3544a14ede7" [ 989.675925] env[62952]: _type = "Task" [ 989.675925] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.683743] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b238c2-21f2-ce2d-bd1e-f3544a14ede7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.818297] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367387, 'name': ReconfigVM_Task, 'duration_secs': 0.99527} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.818597] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 303c9fe7-0d89-4a1a-adcc-6f787486df61/303c9fe7-0d89-4a1a-adcc-6f787486df61.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.819262] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4cb4c3b0-3369-4d30-bd22-6c6e61d1db53 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.826539] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 989.826539] env[62952]: value = "task-1367388" [ 989.826539] env[62952]: _type = "Task" [ 989.826539] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.835615] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367388, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.886112] env[62952]: DEBUG nova.network.neutron [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Successfully created port: 97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.973259] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 990.014275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.066184] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-da749732-72d8-4180-aedf-73a17fa9dea6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.066329] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-da749732-72d8-4180-aedf-73a17fa9dea6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.066475] env[62952]: DEBUG nova.network.neutron [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.186367] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b238c2-21f2-ce2d-bd1e-f3544a14ede7, 'name': SearchDatastore_Task, 'duration_secs': 0.009041} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.189562] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fa604a6-195f-4e37-97dc-e07647c34230 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.194953] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 990.194953] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ed619c-1cd4-8992-472f-8bc49934c5d7" [ 990.194953] env[62952]: _type = "Task" [ 990.194953] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.203080] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ed619c-1cd4-8992-472f-8bc49934c5d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.232099] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06085d58-75af-46ee-8efc-45a3aab631b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.239601] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314fe0d0-6d5e-4b28-b842-6d66177aaa8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.270491] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadc6864-ed3f-444b-9f28-9456a600d7e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.278662] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b80d5d5-a196-45ce-9ce6-a9a6a7cc2c97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.292716] env[62952]: DEBUG nova.compute.provider_tree [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.337317] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367388, 'name': Rename_Task, 'duration_secs': 0.318007} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.338514] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.339607] env[62952]: DEBUG nova.compute.manager [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Received event network-changed-3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.339799] env[62952]: DEBUG nova.compute.manager [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Refreshing instance network info cache due to event network-changed-3582454d-1ad8-4562-8f55-42a3f083d328. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.339992] env[62952]: DEBUG oslo_concurrency.lockutils [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] Acquiring lock "refresh_cache-da749732-72d8-4180-aedf-73a17fa9dea6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.340201] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f3cd83e-0066-41d8-93ee-eb69f3df6045 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.348095] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 990.348095] env[62952]: value = "task-1367389" [ 990.348095] env[62952]: _type = "Task" [ 990.348095] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.354124] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.552855] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 990.553218] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290942', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'name': 'volume-5b018b5e-81cf-4820-85bd-d3015cbda494', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b', 'attached_at': '', 'detached_at': '', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'serial': '5b018b5e-81cf-4820-85bd-d3015cbda494'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 990.554115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b1a066-e9d8-4b3e-8e55-51433c5bd6ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.572933] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea1100f-0651-4799-8e45-dc0143159032 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.599111] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] volume-5b018b5e-81cf-4820-85bd-d3015cbda494/volume-5b018b5e-81cf-4820-85bd-d3015cbda494.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.599442] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-085b97b1-512e-4210-84dd-6887c51c91fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.613677] env[62952]: DEBUG nova.network.neutron [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 990.622127] env[62952]: DEBUG oslo_vmware.api [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 990.622127] env[62952]: value = "task-1367390" [ 990.622127] env[62952]: _type = "Task" [ 990.622127] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.631235] env[62952]: DEBUG oslo_vmware.api [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367390, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.704908] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ed619c-1cd4-8992-472f-8bc49934c5d7, 'name': SearchDatastore_Task, 'duration_secs': 0.010248} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.705253] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.705533] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 244f7b73-fda4-4477-b04d-c4f5cedd2d98/244f7b73-fda4-4477-b04d-c4f5cedd2d98.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.705807] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d78bc455-1c8e-4cc9-8481-e923ab20212b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.714798] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 990.714798] env[62952]: value = "task-1367391" [ 990.714798] env[62952]: _type = "Task" [ 990.714798] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.722774] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.749518] env[62952]: DEBUG nova.network.neutron [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Updating instance_info_cache with network_info: [{"id": "3582454d-1ad8-4562-8f55-42a3f083d328", "address": "fa:16:3e:82:62:d5", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3582454d-1a", "ovs_interfaceid": "3582454d-1ad8-4562-8f55-42a3f083d328", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.795667] env[62952]: DEBUG nova.scheduler.client.report [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.857714] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367389, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.971837] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "97995e38-b27e-478a-8553-eb1c844bb0a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.972247] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "97995e38-b27e-478a-8553-eb1c844bb0a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.972529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "97995e38-b27e-478a-8553-eb1c844bb0a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.972750] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "97995e38-b27e-478a-8553-eb1c844bb0a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.972934] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "97995e38-b27e-478a-8553-eb1c844bb0a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.975916] env[62952]: INFO nova.compute.manager [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Terminating instance [ 990.978176] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "refresh_cache-97995e38-b27e-478a-8553-eb1c844bb0a6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.978347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "refresh_cache-97995e38-b27e-478a-8553-eb1c844bb0a6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.978550] env[62952]: DEBUG nova.network.neutron [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.986172] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 991.016729] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.017149] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.017417] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.017734] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.017945] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.018209] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.018680] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.018773] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.019059] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.019338] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.019613] env[62952]: DEBUG nova.virt.hardware [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.020916] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7da896-020a-43be-9c70-9839cc045b05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.032929] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82acd11e-fe25-4740-b34e-43c7b4a91c77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.133527] env[62952]: DEBUG oslo_vmware.api [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367390, 'name': ReconfigVM_Task, 'duration_secs': 0.409404} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.133940] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Reconfigured VM instance instance-00000046 to attach disk [datastore1] volume-5b018b5e-81cf-4820-85bd-d3015cbda494/volume-5b018b5e-81cf-4820-85bd-d3015cbda494.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.139641] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4995d548-d694-4226-88a5-dbcc00af3469 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.157729] env[62952]: DEBUG oslo_vmware.api [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 991.157729] env[62952]: value = "task-1367392" [ 991.157729] env[62952]: _type = "Task" [ 991.157729] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.166211] env[62952]: DEBUG oslo_vmware.api [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367392, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.225425] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442831} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.225745] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 244f7b73-fda4-4477-b04d-c4f5cedd2d98/244f7b73-fda4-4477-b04d-c4f5cedd2d98.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.225969] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.226251] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-465379b5-fef8-4a93-90a8-b736daefdcc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.234035] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 991.234035] env[62952]: value = "task-1367393" [ 991.234035] env[62952]: _type = "Task" [ 991.234035] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.242269] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.253381] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-da749732-72d8-4180-aedf-73a17fa9dea6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.253842] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Instance network_info: |[{"id": "3582454d-1ad8-4562-8f55-42a3f083d328", "address": "fa:16:3e:82:62:d5", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3582454d-1a", "ovs_interfaceid": "3582454d-1ad8-4562-8f55-42a3f083d328", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.254043] env[62952]: DEBUG oslo_concurrency.lockutils [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] Acquired lock "refresh_cache-da749732-72d8-4180-aedf-73a17fa9dea6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.254238] env[62952]: DEBUG nova.network.neutron [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Refreshing network info cache for port 3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.255548] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:62:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3582454d-1ad8-4562-8f55-42a3f083d328', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.263369] env[62952]: DEBUG oslo.service.loopingcall [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.264366] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.264602] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae21fa31-85e3-4ac5-bc22-1636208a5db2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.285113] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.285113] env[62952]: value = "task-1367394" [ 991.285113] env[62952]: _type = "Task" [ 991.285113] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.295190] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367394, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.300972] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.301946] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.304686] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.511s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.306149] env[62952]: INFO nova.compute.claims [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.357815] env[62952]: DEBUG oslo_vmware.api [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367389, 'name': PowerOnVM_Task, 'duration_secs': 0.515195} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.358110] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.358322] env[62952]: INFO nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Took 7.47 seconds to spawn the instance on the hypervisor. [ 991.358506] env[62952]: DEBUG nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.359415] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc0dea2-cded-43cd-b005-3ba701e6a5a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.470903] env[62952]: DEBUG nova.network.neutron [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Successfully updated port: 97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.502329] env[62952]: DEBUG nova.network.neutron [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.548808] env[62952]: DEBUG nova.network.neutron [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.667704] env[62952]: DEBUG oslo_vmware.api [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367392, 'name': ReconfigVM_Task, 'duration_secs': 0.253144} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.668061] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290942', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'name': 'volume-5b018b5e-81cf-4820-85bd-d3015cbda494', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b', 'attached_at': '', 'detached_at': '', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'serial': '5b018b5e-81cf-4820-85bd-d3015cbda494'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 991.746634] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071166} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.746940] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.747781] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdb6fe7-22b9-4d68-a210-2a67432e6e31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.771406] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 244f7b73-fda4-4477-b04d-c4f5cedd2d98/244f7b73-fda4-4477-b04d-c4f5cedd2d98.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.774765] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ad2064c-d2f5-4a09-90db-fa0d540eda0f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.802294] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367394, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.804119] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 991.804119] env[62952]: value = "task-1367395" [ 991.804119] env[62952]: _type = "Task" [ 991.804119] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.813063] env[62952]: DEBUG nova.compute.utils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.817052] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367395, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.817205] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.817392] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.891353] env[62952]: DEBUG nova.policy [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '378e043e429e4b3098e0003311d0b6de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f7b88e9cabf41a7802ebd6b3b84d51e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.894860] env[62952]: INFO nova.compute.manager [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Took 27.66 seconds to build instance. [ 991.974605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-1e02f16d-b86e-4533-928b-a672ce69788f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.974730] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-1e02f16d-b86e-4533-928b-a672ce69788f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.974977] env[62952]: DEBUG nova.network.neutron [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.052102] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "refresh_cache-97995e38-b27e-478a-8553-eb1c844bb0a6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.052637] env[62952]: DEBUG nova.compute.manager [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.052770] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.054203] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7120b34a-9503-4cf8-9d58-cc65a3b17496 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.063769] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.063769] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-682a9844-26bd-4237-92f7-ead0669c4c4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.069828] env[62952]: DEBUG oslo_vmware.api [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 992.069828] env[62952]: value = "task-1367396" [ 992.069828] env[62952]: _type = "Task" [ 992.069828] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.078488] env[62952]: DEBUG oslo_vmware.api [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.303424] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367394, 'name': CreateVM_Task, 'duration_secs': 0.612477} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.303646] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.304300] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.304459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.304774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.305117] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11908251-3f2f-45c0-86da-ebba69f3de26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.313367] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 992.313367] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52db45d6-f551-d1d0-f171-405ed0cc20f3" [ 992.313367] env[62952]: _type = "Task" [ 992.313367] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.317537] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367395, 'name': ReconfigVM_Task, 'duration_secs': 0.293969} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.321029] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 244f7b73-fda4-4477-b04d-c4f5cedd2d98/244f7b73-fda4-4477-b04d-c4f5cedd2d98.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.321463] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.326922] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cf1b55b-5747-41fd-a16d-f1ad84adff8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.335884] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52db45d6-f551-d1d0-f171-405ed0cc20f3, 'name': SearchDatastore_Task, 'duration_secs': 0.009731} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.336543] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.336914] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.337266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.337518] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.337792] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.338194] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 992.338194] env[62952]: value = "task-1367397" [ 992.338194] env[62952]: _type = "Task" [ 992.338194] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.338533] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd29ad36-c1ab-4639-8560-1585e77c7235 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.342332] env[62952]: DEBUG nova.network.neutron [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Updated VIF entry in instance network info cache for port 3582454d-1ad8-4562-8f55-42a3f083d328. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.342781] env[62952]: DEBUG nova.network.neutron [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Updating instance_info_cache with network_info: [{"id": "3582454d-1ad8-4562-8f55-42a3f083d328", "address": "fa:16:3e:82:62:d5", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3582454d-1a", "ovs_interfaceid": "3582454d-1ad8-4562-8f55-42a3f083d328", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.357812] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367397, 'name': Rename_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.359910] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.359910] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.359910] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf0b2545-66b3-4542-8b95-5acd8f6a26ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.366672] env[62952]: DEBUG nova.compute.manager [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Received event network-vif-plugged-97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.366672] env[62952]: DEBUG oslo_concurrency.lockutils [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] Acquiring lock "1e02f16d-b86e-4533-928b-a672ce69788f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.366979] env[62952]: DEBUG oslo_concurrency.lockutils [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] Lock "1e02f16d-b86e-4533-928b-a672ce69788f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.367260] env[62952]: DEBUG oslo_concurrency.lockutils [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] Lock "1e02f16d-b86e-4533-928b-a672ce69788f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.367519] env[62952]: DEBUG nova.compute.manager [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] No waiting events found dispatching network-vif-plugged-97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.367764] env[62952]: WARNING nova.compute.manager [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Received unexpected event network-vif-plugged-97ceb54a-06d4-4154-9965-95093341cdb4 for instance with vm_state building and task_state spawning. [ 992.367990] env[62952]: DEBUG nova.compute.manager [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Received event network-changed-97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.368246] env[62952]: DEBUG nova.compute.manager [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Refreshing instance network info cache due to event network-changed-97ceb54a-06d4-4154-9965-95093341cdb4. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.368481] env[62952]: DEBUG oslo_concurrency.lockutils [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] Acquiring lock "refresh_cache-1e02f16d-b86e-4533-928b-a672ce69788f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.372720] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 992.372720] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524c36c9-33c1-7abb-0b1a-3fabd6fd8424" [ 992.372720] env[62952]: _type = "Task" [ 992.372720] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.385682] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524c36c9-33c1-7abb-0b1a-3fabd6fd8424, 'name': SearchDatastore_Task, 'duration_secs': 0.010102} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.388932] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70bf0a53-f63c-400a-ab58-f6cc1bd96fab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.394477] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 992.394477] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52688fda-15b0-6b8d-1ef7-88cbf4d59a19" [ 992.394477] env[62952]: _type = "Task" [ 992.394477] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.398065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-67b63e79-e914-43d3-8c7d-497d7ac6a20d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.182s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.402856] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52688fda-15b0-6b8d-1ef7-88cbf4d59a19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.527538] env[62952]: DEBUG nova.network.neutron [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 992.548251] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Successfully created port: 567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.579894] env[62952]: DEBUG oslo_vmware.api [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367396, 'name': PowerOffVM_Task, 'duration_secs': 0.259089} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.582313] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.582585] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.583055] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bdabea5-4826-4178-83b0-61c66bad949b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.592738] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b128a7-022b-41f8-a134-9a829adae37d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.602012] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d597879-8f16-4322-80b0-a9c09f4d2054 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.607958] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.609026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.609026] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleting the datastore file [datastore2] 97995e38-b27e-478a-8553-eb1c844bb0a6 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.631809] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fad6359e-41a1-4130-9558-d6c0c4549008 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.634585] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a94465f-e2c7-4517-8a5f-452a00b87104 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.643717] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffe0ea9-e26a-450a-9694-001126640d27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.648196] env[62952]: DEBUG oslo_vmware.api [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 992.648196] env[62952]: value = "task-1367399" [ 992.648196] env[62952]: _type = "Task" [ 992.648196] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.661747] env[62952]: DEBUG nova.compute.provider_tree [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.670166] env[62952]: DEBUG oslo_vmware.api [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367399, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.710503] env[62952]: DEBUG nova.objects.instance [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.813510] env[62952]: DEBUG nova.network.neutron [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Updating instance_info_cache with network_info: [{"id": "97ceb54a-06d4-4154-9965-95093341cdb4", "address": "fa:16:3e:f0:e3:25", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97ceb54a-06", "ovs_interfaceid": "97ceb54a-06d4-4154-9965-95093341cdb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.853414] env[62952]: DEBUG oslo_concurrency.lockutils [req-330847d4-4a3c-46b0-9248-5c5b9ee57469 req-30c00fa0-ce59-463f-90e7-df406ffc696c service nova] Releasing lock "refresh_cache-da749732-72d8-4180-aedf-73a17fa9dea6" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.855240] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367397, 'name': Rename_Task, 'duration_secs': 0.157049} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.855780] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.856050] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8baa546e-4a4f-44cf-b721-5403629440c4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.865229] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 992.865229] env[62952]: value = "task-1367400" [ 992.865229] env[62952]: _type = "Task" [ 992.865229] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.873728] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.905380] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52688fda-15b0-6b8d-1ef7-88cbf4d59a19, 'name': SearchDatastore_Task, 'duration_secs': 0.011717} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.905380] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.905380] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] da749732-72d8-4180-aedf-73a17fa9dea6/da749732-72d8-4180-aedf-73a17fa9dea6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.905380] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3d320e7-f18b-46c2-9bf8-fa04aee64fb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.911826] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 992.911826] env[62952]: value = "task-1367401" [ 992.911826] env[62952]: _type = "Task" [ 992.911826] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.919762] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367401, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.161580] env[62952]: DEBUG oslo_vmware.api [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367399, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276147} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.161580] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.161580] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 993.161580] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 993.162048] env[62952]: INFO nova.compute.manager [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 993.162386] env[62952]: DEBUG oslo.service.loopingcall [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.162610] env[62952]: DEBUG nova.compute.manager [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 993.162702] env[62952]: DEBUG nova.network.neutron [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 993.168549] env[62952]: DEBUG nova.scheduler.client.report [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.189400] env[62952]: DEBUG nova.network.neutron [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.214993] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8b1dc091-1e3f-44a8-ac0f-c6f8bbc0d34c tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.269s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.322960] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-1e02f16d-b86e-4533-928b-a672ce69788f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.323441] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance network_info: |[{"id": "97ceb54a-06d4-4154-9965-95093341cdb4", "address": "fa:16:3e:f0:e3:25", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97ceb54a-06", "ovs_interfaceid": "97ceb54a-06d4-4154-9965-95093341cdb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 993.323846] env[62952]: DEBUG oslo_concurrency.lockutils [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] Acquired lock "refresh_cache-1e02f16d-b86e-4533-928b-a672ce69788f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.324978] env[62952]: DEBUG nova.network.neutron [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Refreshing network info cache for port 97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.329158] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:e3:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97ceb54a-06d4-4154-9965-95093341cdb4', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.334614] env[62952]: DEBUG oslo.service.loopingcall [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.335248] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 993.336872] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a94967a-8265-4671-b72a-6dc8194d5b44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.354511] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.364695] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.364695] env[62952]: value = "task-1367402" [ 993.364695] env[62952]: _type = "Task" [ 993.364695] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.379712] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367402, 'name': CreateVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.385341] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367400, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.394406] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.394684] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.394871] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.395322] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.395412] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.395565] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.395795] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.395968] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.396176] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.396355] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.396540] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.397898] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2e74c7-7a47-4ba8-80eb-333512267195 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.408216] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738af4a2-8d0c-424b-9eb7-b2f63ea250c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.429245] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367401, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.435235] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.435522] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.435712] env[62952]: DEBUG nova.compute.manager [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.437105] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb15a61-54cd-4e48-8907-1044701e36ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.444216] env[62952]: DEBUG nova.compute.manager [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 993.444896] env[62952]: DEBUG nova.objects.instance [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.560511] env[62952]: DEBUG nova.compute.manager [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-changed-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.560651] env[62952]: DEBUG nova.compute.manager [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing instance network info cache due to event network-changed-48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.560868] env[62952]: DEBUG oslo_concurrency.lockutils [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.561020] env[62952]: DEBUG oslo_concurrency.lockutils [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.561198] env[62952]: DEBUG nova.network.neutron [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.675521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.676008] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.679386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.043s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.681046] env[62952]: INFO nova.compute.claims [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.695912] env[62952]: DEBUG nova.network.neutron [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.880502] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367402, 'name': CreateVM_Task, 'duration_secs': 0.467403} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.883474] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.884580] env[62952]: DEBUG oslo_vmware.api [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367400, 'name': PowerOnVM_Task, 'duration_secs': 0.753337} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.886476] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.889025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.889025] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.889025] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.889025] env[62952]: INFO nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Took 7.64 seconds to spawn the instance on the hypervisor. [ 993.889025] env[62952]: DEBUG nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.889375] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-363456b9-4747-424d-ba7a-c7b8bc655b6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.890363] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36328f2-43a8-4b90-87fe-89da1a9f8994 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.897389] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 993.897389] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ac363a-8232-a088-a383-e4312c822ad0" [ 993.897389] env[62952]: _type = "Task" [ 993.897389] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.909817] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ac363a-8232-a088-a383-e4312c822ad0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.921224] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367401, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530367} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.924022] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] da749732-72d8-4180-aedf-73a17fa9dea6/da749732-72d8-4180-aedf-73a17fa9dea6.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.924022] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.924022] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db6151a4-801f-4750-a460-c5652a3d118d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.928429] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 993.928429] env[62952]: value = "task-1367403" [ 993.928429] env[62952]: _type = "Task" [ 993.928429] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.938570] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367403, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.950756] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.951024] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c21293a1-e718-4556-80ea-dfb494b11eab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.959462] env[62952]: DEBUG oslo_vmware.api [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 993.959462] env[62952]: value = "task-1367404" [ 993.959462] env[62952]: _type = "Task" [ 993.959462] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.970805] env[62952]: DEBUG oslo_vmware.api [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.185659] env[62952]: DEBUG nova.compute.utils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 994.189155] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 994.189345] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 994.194191] env[62952]: DEBUG nova.network.neutron [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Updated VIF entry in instance network info cache for port 97ceb54a-06d4-4154-9965-95093341cdb4. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.196333] env[62952]: DEBUG nova.network.neutron [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Updating instance_info_cache with network_info: [{"id": "97ceb54a-06d4-4154-9965-95093341cdb4", "address": "fa:16:3e:f0:e3:25", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97ceb54a-06", "ovs_interfaceid": "97ceb54a-06d4-4154-9965-95093341cdb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.198479] env[62952]: INFO nova.compute.manager [-] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Took 1.04 seconds to deallocate network for instance. [ 994.291255] env[62952]: DEBUG nova.policy [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '378e043e429e4b3098e0003311d0b6de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f7b88e9cabf41a7802ebd6b3b84d51e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 994.420453] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ac363a-8232-a088-a383-e4312c822ad0, 'name': SearchDatastore_Task, 'duration_secs': 0.041125} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.420504] env[62952]: INFO nova.compute.manager [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Took 29.48 seconds to build instance. [ 994.422513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.422513] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.422513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.422912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.423169] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.423826] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4d4196b-4fe8-46fe-8ebb-1ceccab5df0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.435051] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.435304] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.439121] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab3ab6ae-6a39-495d-be69-6cf09c55f77d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.441593] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367403, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117175} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.442523] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.443923] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aca06ad-4886-472b-a6b9-3bb30578c659 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.447544] env[62952]: DEBUG nova.network.neutron [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updated VIF entry in instance network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.447912] env[62952]: DEBUG nova.network.neutron [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.450819] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 994.450819] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5264beb8-44fc-10b6-c4e1-2c5cf4120af7" [ 994.450819] env[62952]: _type = "Task" [ 994.450819] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.481436] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] da749732-72d8-4180-aedf-73a17fa9dea6/da749732-72d8-4180-aedf-73a17fa9dea6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.481436] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Successfully updated port: 567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 994.486734] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b69fc56e-22c4-4c0b-8a33-bc98d31ea28e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.507059] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5264beb8-44fc-10b6-c4e1-2c5cf4120af7, 'name': SearchDatastore_Task, 'duration_secs': 0.035398} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.508633] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b129691-e6e3-4d51-824b-2baa61383bc1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.514097] env[62952]: DEBUG oslo_vmware.api [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367404, 'name': PowerOffVM_Task, 'duration_secs': 0.53671} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.515856] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.516061] env[62952]: DEBUG nova.compute.manager [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.516468] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 994.516468] env[62952]: value = "task-1367405" [ 994.516468] env[62952]: _type = "Task" [ 994.516468] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.517245] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63f2b7a-17b2-432a-a6f6-3d15e0b81a39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.524232] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 994.524232] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52aba64e-cf00-0e92-ec20-8b7f93e4f223" [ 994.524232] env[62952]: _type = "Task" [ 994.524232] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.537375] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367405, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.539842] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52aba64e-cf00-0e92-ec20-8b7f93e4f223, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.693022] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.699337] env[62952]: DEBUG oslo_concurrency.lockutils [req-bdbb3dda-17aa-4093-a9b2-e1aef0898232 req-44888bc2-88e0-4f07-88cb-3cb79c83ac58 service nova] Releasing lock "refresh_cache-1e02f16d-b86e-4533-928b-a672ce69788f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.705332] env[62952]: DEBUG nova.compute.manager [req-055a401e-eb35-4a0b-bbd8-763e2dda2de0 req-ba13ca07-0062-4528-b9c0-01b9494b7c46 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Received event network-vif-plugged-567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.705598] env[62952]: DEBUG oslo_concurrency.lockutils [req-055a401e-eb35-4a0b-bbd8-763e2dda2de0 req-ba13ca07-0062-4528-b9c0-01b9494b7c46 service nova] Acquiring lock "4c4845e6-79ca-4a2a-a636-842d6919e70f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.705861] env[62952]: DEBUG oslo_concurrency.lockutils [req-055a401e-eb35-4a0b-bbd8-763e2dda2de0 req-ba13ca07-0062-4528-b9c0-01b9494b7c46 service nova] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.706056] env[62952]: DEBUG oslo_concurrency.lockutils [req-055a401e-eb35-4a0b-bbd8-763e2dda2de0 req-ba13ca07-0062-4528-b9c0-01b9494b7c46 service nova] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.706241] env[62952]: DEBUG nova.compute.manager [req-055a401e-eb35-4a0b-bbd8-763e2dda2de0 req-ba13ca07-0062-4528-b9c0-01b9494b7c46 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] No waiting events found dispatching network-vif-plugged-567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.706420] env[62952]: WARNING nova.compute.manager [req-055a401e-eb35-4a0b-bbd8-763e2dda2de0 req-ba13ca07-0062-4528-b9c0-01b9494b7c46 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Received unexpected event network-vif-plugged-567af111-f74a-4736-b704-de3514ef2899 for instance with vm_state building and task_state spawning. [ 994.709030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.852135] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Successfully created port: cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.925964] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f48a24c9-ccd3-46d0-8e36-2b3b629d7e99 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.989s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.952531] env[62952]: DEBUG oslo_concurrency.lockutils [req-cfc4620e-5e6e-4a8e-826d-dc61e6bf1a2b req-b41b24bd-082c-49c5-b10d-3157409eb981 service nova] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.973318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d656b76-f6df-43b9-9d97-eaad27d1bf6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.984059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195f066d-9f74-4365-816b-75d234554b82 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.017728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "refresh_cache-4c4845e6-79ca-4a2a-a636-842d6919e70f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.017728] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "refresh_cache-4c4845e6-79ca-4a2a-a636-842d6919e70f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.017728] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.019324] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cf9d66-9a9a-4d29-9424-17c6e3a2580e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.034746] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367405, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.039160] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18af57f8-f8f7-47b1-91a5-d51fb9384102 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.050127] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52aba64e-cf00-0e92-ec20-8b7f93e4f223, 'name': SearchDatastore_Task, 'duration_secs': 0.02307} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.050605] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bd90628c-da57-452a-b64e-dd89832f7081 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.615s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.052542] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.054063] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.054063] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6429eab9-fa4a-4637-85cf-18577b3d2a07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.065754] env[62952]: DEBUG nova.compute.provider_tree [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.077207] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 995.077207] env[62952]: value = "task-1367406" [ 995.077207] env[62952]: _type = "Task" [ 995.077207] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.087440] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.532455] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367405, 'name': ReconfigVM_Task, 'duration_secs': 0.781111} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.532882] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Reconfigured VM instance instance-00000053 to attach disk [datastore1] da749732-72d8-4180-aedf-73a17fa9dea6/da749732-72d8-4180-aedf-73a17fa9dea6.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.533858] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddc6171b-7653-4f64-a23c-c97ed082e4bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.541576] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 995.541576] env[62952]: value = "task-1367407" [ 995.541576] env[62952]: _type = "Task" [ 995.541576] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.556381] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367407, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.569930] env[62952]: DEBUG nova.scheduler.client.report [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.576607] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.588094] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367406, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.678213] env[62952]: DEBUG nova.compute.manager [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.678505] env[62952]: DEBUG nova.compute.manager [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing instance network info cache due to event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.678776] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.678952] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.680348] env[62952]: DEBUG nova.network.neutron [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.704391] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.706983] env[62952]: DEBUG nova.compute.manager [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.708380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033efd90-08d4-4e0c-9375-cac854fd4432 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.734086] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.734706] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.735067] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.735238] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.735458] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.735659] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.735906] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.736116] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.736410] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.736532] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.736743] env[62952]: DEBUG nova.virt.hardware [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.737724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77315a3-422c-491e-93ef-15f33cbcaf1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.748118] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac36683-38ea-4eb9-97c2-56b94b0f945c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.818166] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Updating instance_info_cache with network_info: [{"id": "567af111-f74a-4736-b704-de3514ef2899", "address": "fa:16:3e:85:aa:65", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567af111-f7", "ovs_interfaceid": "567af111-f74a-4736-b704-de3514ef2899", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.051778] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367407, 'name': Rename_Task, 'duration_secs': 0.212867} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.052125] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 996.052421] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23abefe3-db1c-48d4-aefb-7c9df3b75665 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.058543] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 996.058543] env[62952]: value = "task-1367408" [ 996.058543] env[62952]: _type = "Task" [ 996.058543] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.065703] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.074798] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.075533] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.080231] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.414s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.080546] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.083204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.440s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.083430] env[62952]: DEBUG nova.objects.instance [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lazy-loading 'resources' on Instance uuid 8b2e4601-b981-48d1-84e6-21da66d5ea5f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.094578] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634882} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.095510] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.095742] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.095999] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a8dfd09-4bf5-4f4a-9595-71abc92b702d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.103224] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 996.103224] env[62952]: value = "task-1367409" [ 996.103224] env[62952]: _type = "Task" [ 996.103224] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.112073] env[62952]: DEBUG nova.compute.manager [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.112420] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367409, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.113149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9792ec-7569-4755-a9a3-3d5255890415 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.219202] env[62952]: INFO nova.compute.manager [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] instance snapshotting [ 996.219843] env[62952]: DEBUG nova.objects.instance [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'flavor' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.324736] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "refresh_cache-4c4845e6-79ca-4a2a-a636-842d6919e70f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.325079] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Instance network_info: |[{"id": "567af111-f74a-4736-b704-de3514ef2899", "address": "fa:16:3e:85:aa:65", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567af111-f7", "ovs_interfaceid": "567af111-f74a-4736-b704-de3514ef2899", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 996.325531] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:aa:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '567af111-f74a-4736-b704-de3514ef2899', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.334181] env[62952]: DEBUG oslo.service.loopingcall [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.334583] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 996.334701] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb6b4500-93dc-4ddb-bf19-aa7337ff324d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.355569] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.355569] env[62952]: value = "task-1367410" [ 996.355569] env[62952]: _type = "Task" [ 996.355569] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.365112] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367410, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.372707] env[62952]: DEBUG nova.objects.instance [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.400166] env[62952]: DEBUG nova.network.neutron [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updated VIF entry in instance network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 996.400437] env[62952]: DEBUG nova.network.neutron [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.571880] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.581464] env[62952]: DEBUG nova.compute.utils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.582878] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 996.583113] env[62952]: DEBUG nova.network.neutron [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.596259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3faeff08-ef32-467f-aa7f-036644897e2e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 45.689s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.596259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 24.691s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.596259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.596259] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.596476] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.598832] env[62952]: INFO nova.compute.manager [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Terminating instance [ 996.600264] env[62952]: DEBUG nova.compute.manager [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.600359] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.600901] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0da8598-3bd4-43a4-9387-6f4e1a6e1361 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.615204] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367409, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093483} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.619440] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.619657] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adc62e3-fc1e-43e0-ab20-548b0059df59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.624984] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc09bad-4cf8-4153-a3f4-18204de57cbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.638925] env[62952]: DEBUG nova.policy [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e67193d02201461bb7e4339b5025ca48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfc390d64c0463190f071f3f62936dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.641065] env[62952]: INFO nova.compute.manager [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] instance snapshotting [ 996.644585] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27abf52a-dd82-4611-adc6-ff646b3f86f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.667130] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.678885] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b65f387-980e-4fa6-8fd5-bfd11b9c153b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.692930] env[62952]: WARNING nova.virt.vmwareapi.vmops [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ace8d6f-49ab-4486-bee1-04cd24f7f430 could not be found. [ 996.693167] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.693353] env[62952]: INFO nova.compute.manager [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Took 0.09 seconds to destroy the instance on the hypervisor. [ 996.693607] env[62952]: DEBUG oslo.service.loopingcall [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.708688] env[62952]: DEBUG nova.compute.manager [-] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.708802] env[62952]: DEBUG nova.network.neutron [-] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.714234] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca146c8-a8f0-4d9c-ad8f-064c43ecea95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.719345] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 996.719345] env[62952]: value = "task-1367411" [ 996.719345] env[62952]: _type = "Task" [ 996.719345] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.736186] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300ff1a3-803c-4b8b-b4e4-b58e0819e165 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.744768] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367411, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.762404] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd935a4a-e2de-4888-b51c-1fd1bf4242dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.867191] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367410, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.879216] env[62952]: DEBUG nova.compute.manager [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Received event network-changed-567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.879421] env[62952]: DEBUG nova.compute.manager [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Refreshing instance network info cache due to event network-changed-567af111-f74a-4736-b704-de3514ef2899. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.879675] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] Acquiring lock "refresh_cache-4c4845e6-79ca-4a2a-a636-842d6919e70f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.879827] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] Acquired lock "refresh_cache-4c4845e6-79ca-4a2a-a636-842d6919e70f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.879994] env[62952]: DEBUG nova.network.neutron [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Refreshing network info cache for port 567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.883531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.883731] env[62952]: DEBUG oslo_concurrency.lockutils [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.883904] env[62952]: DEBUG nova.network.neutron [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.884101] env[62952]: DEBUG nova.objects.instance [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'info_cache' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.903063] env[62952]: DEBUG oslo_concurrency.lockutils [req-f9dfae91-d4ae-415d-b250-50114e7576bb req-60bc79be-e3c2-4418-8cc3-669bcd0d1144 service nova] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.961405] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20b6877-ae54-4d53-b32e-51f36c07135d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.969335] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6549dcd-c745-4285-8683-cfef32925b94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.014977] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255edbf6-2050-4dde-ae1c-14f96469e893 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.024236] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a3ffe0-0cd8-44d9-ae17-ce2fb2306016 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.042070] env[62952]: DEBUG nova.compute.provider_tree [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.049537] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Successfully updated port: cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.070081] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.086393] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.102348] env[62952]: DEBUG nova.network.neutron [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Successfully created port: 6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.231901] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367411, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.233948] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 997.234240] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d8c19b69-56d4-4c74-87a1-0da0638846b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.241179] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 997.241179] env[62952]: value = "task-1367412" [ 997.241179] env[62952]: _type = "Task" [ 997.241179] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.250592] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367412, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.276121] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 997.276298] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d05e457c-e4c6-423b-b9c3-cb4a6fbac578 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.284270] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 997.284270] env[62952]: value = "task-1367413" [ 997.284270] env[62952]: _type = "Task" [ 997.284270] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.293243] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367413, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.365559] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367410, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.388133] env[62952]: DEBUG nova.objects.base [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Object Instance<88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b> lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 997.546841] env[62952]: DEBUG nova.scheduler.client.report [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.550198] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "interface-ff189c9e-9e7c-4217-9c65-0f821393870a-e1654a86-c311-45dd-9cc6-143cf350df90" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.550454] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-ff189c9e-9e7c-4217-9c65-0f821393870a-e1654a86-c311-45dd-9cc6-143cf350df90" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.550828] env[62952]: DEBUG nova.objects.instance [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lazy-loading 'flavor' on Instance uuid ff189c9e-9e7c-4217-9c65-0f821393870a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.552304] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "refresh_cache-ba2dfb85-0ffa-4883-abf1-482441c5afb8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.552433] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "refresh_cache-ba2dfb85-0ffa-4883-abf1-482441c5afb8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.552541] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.570582] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.617812] env[62952]: DEBUG nova.network.neutron [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Updated VIF entry in instance network info cache for port 567af111-f74a-4736-b704-de3514ef2899. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 997.618186] env[62952]: DEBUG nova.network.neutron [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Updating instance_info_cache with network_info: [{"id": "567af111-f74a-4736-b704-de3514ef2899", "address": "fa:16:3e:85:aa:65", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567af111-f7", "ovs_interfaceid": "567af111-f74a-4736-b704-de3514ef2899", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.732402] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367411, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.749729] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367412, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.777210] env[62952]: DEBUG nova.network.neutron [-] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.794573] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367413, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.797220] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.797416] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing instance network info cache due to event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.797651] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.797803] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.797966] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.866697] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367410, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.053256] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.055643] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.880s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.055939] env[62952]: DEBUG nova.objects.instance [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 998.073313] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.079295] env[62952]: INFO nova.scheduler.client.report [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Deleted allocations for instance 8b2e4601-b981-48d1-84e6-21da66d5ea5f [ 998.096357] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.118066] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 998.124043] env[62952]: DEBUG oslo_concurrency.lockutils [req-1b6e7e3f-4eb5-42db-95a9-9cefe6eb011d req-d70e3782-bfd9-498d-b74b-35de67a8adc5 service nova] Releasing lock "refresh_cache-4c4845e6-79ca-4a2a-a636-842d6919e70f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.127543] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.128215] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.128602] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.128873] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.129172] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.129341] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.129556] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.129724] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.129895] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.130075] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.130258] env[62952]: DEBUG nova.virt.hardware [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.131380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80aa32c-b0ca-4b0f-bbba-64f1926aba69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.142445] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d789733e-6141-4c01-9016-b93a82ae7623 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.192062] env[62952]: DEBUG nova.objects.instance [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lazy-loading 'pci_requests' on Instance uuid ff189c9e-9e7c-4217-9c65-0f821393870a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.202169] env[62952]: DEBUG nova.network.neutron [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [{"id": "aa4656a7-42d8-47ba-a29d-817000a4b596", "address": "fa:16:3e:14:80:1d", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa4656a7-42", "ovs_interfaceid": "aa4656a7-42d8-47ba-a29d-817000a4b596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.235396] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367411, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.250564] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367412, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.280441] env[62952]: INFO nova.compute.manager [-] [instance: 5ace8d6f-49ab-4486-bee1-04cd24f7f430] Took 1.57 seconds to deallocate network for instance. [ 998.294787] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367413, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.318083] env[62952]: DEBUG nova.network.neutron [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Updating instance_info_cache with network_info: [{"id": "cfd7dd45-fd8c-45d4-a276-ed89144b694f", "address": "fa:16:3e:b8:48:d5", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfd7dd45-fd", "ovs_interfaceid": "cfd7dd45-fd8c-45d4-a276-ed89144b694f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.366633] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367410, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.577925] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.595049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-908f57cc-3bec-4ab9-ba41-ae200ac9ef92 tempest-ServersNegativeTestMultiTenantJSON-1507052375 tempest-ServersNegativeTestMultiTenantJSON-1507052375-project-member] Lock "8b2e4601-b981-48d1-84e6-21da66d5ea5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.018s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.604992] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updated VIF entry in instance network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.605432] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.695023] env[62952]: DEBUG nova.objects.base [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.695366] env[62952]: DEBUG nova.network.neutron [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.704642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.734643] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367411, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.755535] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367412, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.798190] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367413, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.799673] env[62952]: DEBUG nova.policy [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb541b4abd9426ea96edcd066c69be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d1f373267464f14ae2c833151821973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.824036] env[62952]: DEBUG nova.network.neutron [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Successfully updated port: 6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.824036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "refresh_cache-ba2dfb85-0ffa-4883-abf1-482441c5afb8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.824233] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Instance network_info: |[{"id": "cfd7dd45-fd8c-45d4-a276-ed89144b694f", "address": "fa:16:3e:b8:48:d5", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfd7dd45-fd", "ovs_interfaceid": "cfd7dd45-fd8c-45d4-a276-ed89144b694f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 998.824304] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:48:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfd7dd45-fd8c-45d4-a276-ed89144b694f', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 998.832066] env[62952]: DEBUG oslo.service.loopingcall [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.832066] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 998.832066] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3269eb03-d736-4e57-a481-1fa099f35318 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.851769] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 998.851769] env[62952]: value = "task-1367414" [ 998.851769] env[62952]: _type = "Task" [ 998.851769] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.872855] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367414, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.877383] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367410, 'name': CreateVM_Task, 'duration_secs': 2.281738} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.877383] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 998.877623] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.877898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.878163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.878677] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a38c8a43-93d3-4860-93ad-3e13696e1b50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.883678] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 998.883678] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a3ca3b-0d88-2f44-ac0d-b99eb50490d2" [ 998.883678] env[62952]: _type = "Task" [ 998.883678] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.892516] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a3ca3b-0d88-2f44-ac0d-b99eb50490d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.075204] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4eef5f07-e74e-4c3b-b740-d2ee34e96e6b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.077073] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.560s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.078644] env[62952]: INFO nova.compute.claims [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.081156] env[62952]: DEBUG oslo_vmware.api [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367408, 'name': PowerOnVM_Task, 'duration_secs': 2.702343} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.081734] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.081933] env[62952]: INFO nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Took 10.45 seconds to spawn the instance on the hypervisor. [ 999.082220] env[62952]: DEBUG nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.083062] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413c7bd1-dd73-44c7-8a2f-65f73e6ff311 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.109700] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.110132] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-changed-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.111679] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing instance network info cache due to event network-changed-48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.111679] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.111679] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.111679] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.123125] env[62952]: DEBUG nova.compute.manager [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Received event network-vif-plugged-6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.123650] env[62952]: DEBUG oslo_concurrency.lockutils [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.123734] env[62952]: DEBUG oslo_concurrency.lockutils [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.123911] env[62952]: DEBUG oslo_concurrency.lockutils [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.125040] env[62952]: DEBUG nova.compute.manager [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] No waiting events found dispatching network-vif-plugged-6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 999.125040] env[62952]: WARNING nova.compute.manager [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Received unexpected event network-vif-plugged-6e4543b9-4601-4109-b4fc-14eee838c035 for instance with vm_state building and task_state spawning. [ 999.125040] env[62952]: DEBUG nova.compute.manager [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Received event network-changed-6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.125040] env[62952]: DEBUG nova.compute.manager [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Refreshing instance network info cache due to event network-changed-6e4543b9-4601-4109-b4fc-14eee838c035. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.125283] env[62952]: DEBUG oslo_concurrency.lockutils [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] Acquiring lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.125365] env[62952]: DEBUG oslo_concurrency.lockutils [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] Acquired lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.125569] env[62952]: DEBUG nova.network.neutron [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Refreshing network info cache for port 6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.208205] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.208451] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1f4aab5-3115-4477-a3b6-9c668f7bda50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.217740] env[62952]: DEBUG oslo_vmware.api [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 999.217740] env[62952]: value = "task-1367415" [ 999.217740] env[62952]: _type = "Task" [ 999.217740] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.226621] env[62952]: DEBUG oslo_vmware.api [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367415, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.234221] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367411, 'name': ReconfigVM_Task, 'duration_secs': 2.125258} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.234522] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.235185] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdcc7598-981c-4904-a752-6cc69b3d0ab5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.241930] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 999.241930] env[62952]: value = "task-1367416" [ 999.241930] env[62952]: _type = "Task" [ 999.241930] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.253912] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367412, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.257061] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367416, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.299363] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367413, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.321147] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fac305f1-472c-473b-84c1-21b7bba5b41d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "5ace8d6f-49ab-4486-bee1-04cd24f7f430" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.725s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.332297] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.362763] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367414, 'name': CreateVM_Task, 'duration_secs': 0.317748} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.362989] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.363671] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.363862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.364228] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.364500] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47388089-9184-4578-808f-37abd402025a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.371200] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 999.371200] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]520c9ca9-be96-42ad-5632-122d846f2dcd" [ 999.371200] env[62952]: _type = "Task" [ 999.371200] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.379101] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520c9ca9-be96-42ad-5632-122d846f2dcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.395188] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a3ca3b-0d88-2f44-ac0d-b99eb50490d2, 'name': SearchDatastore_Task, 'duration_secs': 0.017437} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.395563] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.395835] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.396133] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.396316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.396535] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.396978] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e52f046-5627-4b83-90a0-e7f0b6418ec1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.405312] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.405515] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 999.406281] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cd109ba-83f1-457d-9e65-4718f39ee690 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.412024] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 999.412024] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ae3ba4-f7e1-e9ff-764c-057747f06856" [ 999.412024] env[62952]: _type = "Task" [ 999.412024] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.420312] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ae3ba4-f7e1-e9ff-764c-057747f06856, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.603779] env[62952]: INFO nova.compute.manager [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Took 32.85 seconds to build instance. [ 999.697483] env[62952]: DEBUG nova.network.neutron [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 999.731409] env[62952]: DEBUG oslo_vmware.api [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367415, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.755725] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367416, 'name': Rename_Task, 'duration_secs': 0.171466} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.760293] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.761293] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367412, 'name': CreateSnapshot_Task, 'duration_secs': 2.147016} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.763821] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac01eba3-ac01-44b7-b5bd-97620f04c356 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.767041] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 999.767041] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530c8c14-126b-44c2-9a62-97e694f2cbb8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.783159] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 999.783159] env[62952]: value = "task-1367417" [ 999.783159] env[62952]: _type = "Task" [ 999.783159] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.799013] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367417, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.804261] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367413, 'name': CreateSnapshot_Task, 'duration_secs': 2.071814} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.804545] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 999.805349] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4f1b22-f593-47cb-8bfe-8d6479e75f97 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.860203] env[62952]: DEBUG nova.network.neutron [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.884163] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520c9ca9-be96-42ad-5632-122d846f2dcd, 'name': SearchDatastore_Task, 'duration_secs': 0.01132} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.884431] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.884702] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.884978] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.885202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.885442] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.885758] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8250d4d0-3340-4706-88fd-24e1f4d71e60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.898289] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.898480] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 999.899218] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a879f0b8-0e8b-40a4-9947-c7e40a229cdd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.902076] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updated VIF entry in instance network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.902424] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.907327] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 999.907327] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52843326-4f3a-8275-5287-098959dcd53e" [ 999.907327] env[62952]: _type = "Task" [ 999.907327] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.918926] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52843326-4f3a-8275-5287-098959dcd53e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.925175] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ae3ba4-f7e1-e9ff-764c-057747f06856, 'name': SearchDatastore_Task, 'duration_secs': 0.012508} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.925863] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b205f58e-6bfd-4d45-aac6-11e7d2ea24ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.931393] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 999.931393] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526b562d-e1bf-32f8-80f3-dcd54a25cd4c" [ 999.931393] env[62952]: _type = "Task" [ 999.931393] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.939776] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526b562d-e1bf-32f8-80f3-dcd54a25cd4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.085097] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.085485] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.105516] env[62952]: DEBUG oslo_concurrency.lockutils [None req-53b718f4-4f55-40d6-8d49-2ac3409e5fb8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "da749732-72d8-4180-aedf-73a17fa9dea6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.370s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.235924] env[62952]: DEBUG oslo_vmware.api [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367415, 'name': PowerOnVM_Task, 'duration_secs': 0.756104} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.236245] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.236502] env[62952]: DEBUG nova.compute.manager [None req-342a38f8-41c1-4f8e-a0c8-06c34f97e132 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.237380] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe0708e-84bf-4c64-a0be-edff730c17e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.291959] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1000.295607] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2084be7f-e58e-430f-ba62-5cc247220fc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.308285] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367417, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.309606] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1000.309606] env[62952]: value = "task-1367418" [ 1000.309606] env[62952]: _type = "Task" [ 1000.309606] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.328041] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1000.329268] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367418, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.329268] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-04953406-1450-4ec8-ad69-c18394f98d51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.345019] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1000.345019] env[62952]: value = "task-1367419" [ 1000.345019] env[62952]: _type = "Task" [ 1000.345019] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.352258] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367419, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.365033] env[62952]: DEBUG oslo_concurrency.lockutils [req-be1e7e0a-dd35-4832-87f0-e060c2620dd9 req-ab9c0165-b57f-4488-a70f-e235654b4c5f service nova] Releasing lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.365033] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.365033] env[62952]: DEBUG nova.network.neutron [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.405525] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.405878] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Received event network-vif-plugged-cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.407205] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquiring lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.407205] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.407205] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.407205] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] No waiting events found dispatching network-vif-plugged-cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.407205] env[62952]: WARNING nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Received unexpected event network-vif-plugged-cfd7dd45-fd8c-45d4-a276-ed89144b694f for instance with vm_state building and task_state spawning. [ 1000.407481] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Received event network-changed-cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.407481] env[62952]: DEBUG nova.compute.manager [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Refreshing instance network info cache due to event network-changed-cfd7dd45-fd8c-45d4-a276-ed89144b694f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.407712] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquiring lock "refresh_cache-ba2dfb85-0ffa-4883-abf1-482441c5afb8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.407882] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Acquired lock "refresh_cache-ba2dfb85-0ffa-4883-abf1-482441c5afb8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.410106] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Refreshing network info cache for port cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.415386] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46fdad3-9d6c-4186-984b-bbd45f173067 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.430591] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52843326-4f3a-8275-5287-098959dcd53e, 'name': SearchDatastore_Task, 'duration_secs': 0.017594} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.434769] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c472ab9f-b948-411d-831d-e93cb9559e4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.442443] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d8ae9b-584a-43b5-8065-2211985bc53b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.455781] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526b562d-e1bf-32f8-80f3-dcd54a25cd4c, 'name': SearchDatastore_Task, 'duration_secs': 0.021338} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.487977] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.488540] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 4c4845e6-79ca-4a2a-a636-842d6919e70f/4c4845e6-79ca-4a2a-a636-842d6919e70f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.488791] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1000.488791] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52dc3142-405d-9d65-cd30-508acb2152dd" [ 1000.488791] env[62952]: _type = "Task" [ 1000.488791] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.490798] env[62952]: DEBUG nova.network.neutron [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Successfully updated port: e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.494433] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02bd7d31-294f-4bc7-a0bf-3fa119d3146b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.494811] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafe7e14-0312-430a-a9df-80d10f5e8a32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.514981] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0140c180-3061-4aa9-990c-612b11a02236 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.519638] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1000.519638] env[62952]: value = "task-1367420" [ 1000.519638] env[62952]: _type = "Task" [ 1000.519638] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.520376] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52dc3142-405d-9d65-cd30-508acb2152dd, 'name': SearchDatastore_Task, 'duration_secs': 0.017501} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.520861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.521278] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] ba2dfb85-0ffa-4883-abf1-482441c5afb8/ba2dfb85-0ffa-4883-abf1-482441c5afb8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.525756] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0604d4a-58e3-4222-bb86-f7547a5d1dfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.537036] env[62952]: DEBUG nova.compute.provider_tree [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.544764] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367420, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.546229] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1000.546229] env[62952]: value = "task-1367421" [ 1000.546229] env[62952]: _type = "Task" [ 1000.546229] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.558283] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.588216] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.695668] env[62952]: DEBUG nova.compute.manager [req-20c59b17-a1d9-4aaf-b1f2-9005e8e4cfdb req-3d866ae0-ef00-479e-8d39-a65303709a4a service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-vif-plugged-e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.696430] env[62952]: DEBUG oslo_concurrency.lockutils [req-20c59b17-a1d9-4aaf-b1f2-9005e8e4cfdb req-3d866ae0-ef00-479e-8d39-a65303709a4a service nova] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.696527] env[62952]: DEBUG oslo_concurrency.lockutils [req-20c59b17-a1d9-4aaf-b1f2-9005e8e4cfdb req-3d866ae0-ef00-479e-8d39-a65303709a4a service nova] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.697061] env[62952]: DEBUG oslo_concurrency.lockutils [req-20c59b17-a1d9-4aaf-b1f2-9005e8e4cfdb req-3d866ae0-ef00-479e-8d39-a65303709a4a service nova] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.697216] env[62952]: DEBUG nova.compute.manager [req-20c59b17-a1d9-4aaf-b1f2-9005e8e4cfdb req-3d866ae0-ef00-479e-8d39-a65303709a4a service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] No waiting events found dispatching network-vif-plugged-e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.697511] env[62952]: WARNING nova.compute.manager [req-20c59b17-a1d9-4aaf-b1f2-9005e8e4cfdb req-3d866ae0-ef00-479e-8d39-a65303709a4a service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received unexpected event network-vif-plugged-e1654a86-c311-45dd-9cc6-143cf350df90 for instance with vm_state active and task_state None. [ 1000.804562] env[62952]: DEBUG oslo_vmware.api [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367417, 'name': PowerOnVM_Task, 'duration_secs': 1.026485} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.804986] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.805682] env[62952]: INFO nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Took 9.82 seconds to spawn the instance on the hypervisor. [ 1000.805682] env[62952]: DEBUG nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.806270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73207600-fd34-4f7d-803d-edf12b5d0e10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.825320] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367418, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.854892] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367419, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.938833] env[62952]: DEBUG nova.network.neutron [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.000481] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.000695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.000873] env[62952]: DEBUG nova.network.neutron [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.036573] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367420, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.045208] env[62952]: DEBUG nova.scheduler.client.report [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.065931] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367421, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.115701] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.200016] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Updated VIF entry in instance network info cache for port cfd7dd45-fd8c-45d4-a276-ed89144b694f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.201210] env[62952]: DEBUG nova.network.neutron [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Updating instance_info_cache with network_info: [{"id": "cfd7dd45-fd8c-45d4-a276-ed89144b694f", "address": "fa:16:3e:b8:48:d5", "network": {"id": "4d45265e-8b9b-4a8a-9522-25047c3083d9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1223551777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f7b88e9cabf41a7802ebd6b3b84d51e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfd7dd45-fd", "ovs_interfaceid": "cfd7dd45-fd8c-45d4-a276-ed89144b694f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.278019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "da749732-72d8-4180-aedf-73a17fa9dea6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.278019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "da749732-72d8-4180-aedf-73a17fa9dea6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.278019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "da749732-72d8-4180-aedf-73a17fa9dea6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.278019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "da749732-72d8-4180-aedf-73a17fa9dea6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.278282] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "da749732-72d8-4180-aedf-73a17fa9dea6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.279147] env[62952]: INFO nova.compute.manager [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Terminating instance [ 1001.281064] env[62952]: DEBUG nova.compute.manager [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.281276] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.284780] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a06af5e-81c5-4cdf-8db1-fde1aefca3dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.293689] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.294287] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efb38234-0e88-4593-a8da-c749f53fc8d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.302309] env[62952]: DEBUG oslo_vmware.api [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1001.302309] env[62952]: value = "task-1367422" [ 1001.302309] env[62952]: _type = "Task" [ 1001.302309] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.331174] env[62952]: DEBUG oslo_vmware.api [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.341708] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367418, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.342817] env[62952]: DEBUG nova.network.neutron [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updating instance_info_cache with network_info: [{"id": "6e4543b9-4601-4109-b4fc-14eee838c035", "address": "fa:16:3e:89:6a:b0", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e4543b9-46", "ovs_interfaceid": "6e4543b9-4601-4109-b4fc-14eee838c035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.344398] env[62952]: INFO nova.compute.manager [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Took 34.50 seconds to build instance. [ 1001.359794] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367419, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.541297] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367420, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.716044} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.541639] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 4c4845e6-79ca-4a2a-a636-842d6919e70f/4c4845e6-79ca-4a2a-a636-842d6919e70f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.541922] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.542246] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0755c804-2219-46d0-a68f-76c4e8ace0d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.549486] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1001.549486] env[62952]: value = "task-1367423" [ 1001.549486] env[62952]: _type = "Task" [ 1001.549486] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.564095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.564751] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1001.567561] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367421, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.571602] env[62952]: WARNING nova.network.neutron [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] cab59c10-e233-4404-bbf5-40841bb12dbd already exists in list: networks containing: ['cab59c10-e233-4404-bbf5-40841bb12dbd']. ignoring it [ 1001.573537] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.595s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.573803] env[62952]: DEBUG nova.objects.instance [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'resources' on Instance uuid 3bd9c356-42d6-450f-9069-b066952d9524 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.574896] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] ba2dfb85-0ffa-4883-abf1-482441c5afb8/ba2dfb85-0ffa-4883-abf1-482441c5afb8.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.575156] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.575436] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.576039] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe44f481-51ad-4067-9c77-fac1140a352f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.583403] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1001.583403] env[62952]: value = "task-1367424" [ 1001.583403] env[62952]: _type = "Task" [ 1001.583403] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.595985] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367424, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.704495] env[62952]: DEBUG oslo_concurrency.lockutils [req-65b1ba67-8d4a-487f-b184-62123417a6e4 req-abffb726-f4c6-446c-93ae-6f05dce42843 service nova] Releasing lock "refresh_cache-ba2dfb85-0ffa-4883-abf1-482441c5afb8" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.814150] env[62952]: DEBUG oslo_vmware.api [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367422, 'name': PowerOffVM_Task, 'duration_secs': 0.467543} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.817080] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.817317] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.817811] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6663e41-d63b-42fb-8061-b914f1beef71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.824895] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367418, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.853795] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4e7b6b39-0485-43aa-9a21-e1b93f7425eb tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.019s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.854493] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.854950] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Instance network_info: |[{"id": "6e4543b9-4601-4109-b4fc-14eee838c035", "address": "fa:16:3e:89:6a:b0", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e4543b9-46", "ovs_interfaceid": "6e4543b9-4601-4109-b4fc-14eee838c035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.855942] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:6a:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e4543b9-4601-4109-b4fc-14eee838c035', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.865102] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Creating folder: Project (8bfc390d64c0463190f071f3f62936dc). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1001.869110] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef9b15e9-4a5b-4c35-99ec-b4d2d77a3d4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.870936] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367419, 'name': CloneVM_Task, 'duration_secs': 1.393827} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.871220] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created linked-clone VM from snapshot [ 1001.872320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e375355-7010-4551-8372-ebc952fee35e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.884441] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Uploading image 497ca4b0-de44-46b4-b59e-8ecf1e60917f {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1001.887624] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Created folder: Project (8bfc390d64c0463190f071f3f62936dc) in parent group-v290852. [ 1001.887973] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Creating folder: Instances. Parent ref: group-v290952. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1001.888133] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52141294-b338-4c0d-ba37-c61d43fb5930 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.905630] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1001.905630] env[62952]: value = "vm-290951" [ 1001.905630] env[62952]: _type = "VirtualMachine" [ 1001.905630] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1001.906087] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-abd03fbd-f517-4825-86c2-bb5563b7c1b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.912981] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease: (returnval){ [ 1001.912981] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5258792e-9b82-5ebd-50fd-c2152aab93fb" [ 1001.912981] env[62952]: _type = "HttpNfcLease" [ 1001.912981] env[62952]: } obtained for exporting VM: (result){ [ 1001.912981] env[62952]: value = "vm-290951" [ 1001.912981] env[62952]: _type = "VirtualMachine" [ 1001.912981] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1001.913855] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the lease: (returnval){ [ 1001.913855] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5258792e-9b82-5ebd-50fd-c2152aab93fb" [ 1001.913855] env[62952]: _type = "HttpNfcLease" [ 1001.913855] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1001.916394] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.916598] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.916781] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore1] da749732-72d8-4180-aedf-73a17fa9dea6 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.924631] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a413ac3f-b658-4d1c-be15-0b4543bac20f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.926185] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Created folder: Instances in parent group-v290952. [ 1001.926453] env[62952]: DEBUG oslo.service.loopingcall [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.927109] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.927726] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5fdbf22c-80e5-4559-b90f-ee1b0ce555a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.946014] env[62952]: DEBUG oslo_vmware.api [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1001.946014] env[62952]: value = "task-1367429" [ 1001.946014] env[62952]: _type = "Task" [ 1001.946014] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.946240] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1001.946240] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5258792e-9b82-5ebd-50fd-c2152aab93fb" [ 1001.946240] env[62952]: _type = "HttpNfcLease" [ 1001.946240] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1001.948812] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1001.948812] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5258792e-9b82-5ebd-50fd-c2152aab93fb" [ 1001.948812] env[62952]: _type = "HttpNfcLease" [ 1001.948812] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1001.953749] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ba6cc4-aa67-4acd-bcf8-c299768ea561 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.955813] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.955813] env[62952]: value = "task-1367430" [ 1001.955813] env[62952]: _type = "Task" [ 1001.955813] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.965021] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f10991-5525-8b0e-4d9c-71d8c9eb6126/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1001.965021] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f10991-5525-8b0e-4d9c-71d8c9eb6126/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1001.968967] env[62952]: DEBUG oslo_vmware.api [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367429, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.036758] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367430, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.064768] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080895} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.065585] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1002.065883] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4790bb64-373d-44a2-a3a0-19f09fa4c134 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.083700] env[62952]: DEBUG nova.compute.utils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1002.096952] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 4c4845e6-79ca-4a2a-a636-842d6919e70f/4c4845e6-79ca-4a2a-a636-842d6919e70f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.099597] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1002.099837] env[62952]: DEBUG nova.network.neutron [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1002.102275] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c3728dd-8c8b-497e-b022-00ff7f4ab25d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.121633] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-de5bf0ed-a2f5-4ee8-8256-49c8408fd20f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.133849] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367424, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.137027] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1002.137747] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1002.137747] env[62952]: value = "task-1367431" [ 1002.137747] env[62952]: _type = "Task" [ 1002.137747] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.138690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bc1e78-b101-4de5-8156-3f78ea490249 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.159053] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.178850] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] ba2dfb85-0ffa-4883-abf1-482441c5afb8/ba2dfb85-0ffa-4883-abf1-482441c5afb8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.187427] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9ffd321-baec-4e58-b7ce-84670b8a2834 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.203719] env[62952]: DEBUG nova.network.neutron [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1654a86-c311-45dd-9cc6-143cf350df90", "address": "fa:16:3e:2d:97:e7", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1654a86-c3", "ovs_interfaceid": "e1654a86-c311-45dd-9cc6-143cf350df90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.207037] env[62952]: DEBUG nova.policy [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17b50c91ccd347e4b45126a8f23c0278', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9657a113032417fb4b97716a772f8ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1002.214487] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1002.214487] env[62952]: value = "task-1367432" [ 1002.214487] env[62952]: _type = "Task" [ 1002.214487] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.224009] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367432, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.328237] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367418, 'name': CloneVM_Task, 'duration_secs': 1.685963} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.328573] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Created linked-clone VM from snapshot [ 1002.329391] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a944cf-7327-457e-99ce-e57dcaf510c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.338081] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Uploading image 1de1bca9-aeab-4b87-85e8-152a9124e537 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1002.384318] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1002.384318] env[62952]: value = "vm-290950" [ 1002.384318] env[62952]: _type = "VirtualMachine" [ 1002.384318] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1002.384318] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-94ad0b65-2d70-43b1-96fb-6629c98df2ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.396182] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lease: (returnval){ [ 1002.396182] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528aa359-c66e-3a30-2c82-e492170f392f" [ 1002.396182] env[62952]: _type = "HttpNfcLease" [ 1002.396182] env[62952]: } obtained for exporting VM: (result){ [ 1002.396182] env[62952]: value = "vm-290950" [ 1002.396182] env[62952]: _type = "VirtualMachine" [ 1002.396182] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1002.398575] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the lease: (returnval){ [ 1002.398575] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528aa359-c66e-3a30-2c82-e492170f392f" [ 1002.398575] env[62952]: _type = "HttpNfcLease" [ 1002.398575] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1002.406892] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1002.406892] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528aa359-c66e-3a30-2c82-e492170f392f" [ 1002.406892] env[62952]: _type = "HttpNfcLease" [ 1002.406892] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1002.465539] env[62952]: DEBUG oslo_vmware.api [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367429, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.341847} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.466092] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.466556] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.466929] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.467218] env[62952]: INFO nova.compute.manager [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1002.467593] env[62952]: DEBUG oslo.service.loopingcall [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.474780] env[62952]: DEBUG nova.compute.manager [-] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.474970] env[62952]: DEBUG nova.network.neutron [-] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.476693] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367430, 'name': CreateVM_Task, 'duration_secs': 0.38694} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.477148] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.478215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.478475] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.478872] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.479210] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5851d7dc-fd5e-4849-b1b7-113503ff99cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.485841] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1002.485841] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5200e5cc-b9ae-844d-717d-c01e879a84e3" [ 1002.485841] env[62952]: _type = "Task" [ 1002.485841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.495800] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5200e5cc-b9ae-844d-717d-c01e879a84e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.519623] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d81ce2a-70ab-4cc2-a1df-fb48042d04b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.529120] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe9de85-be4e-4a70-8dc5-6ecafee5186c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.569107] env[62952]: DEBUG nova.network.neutron [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Successfully created port: 3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.572594] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc15d4ab-b584-4b43-97bf-615c84df4931 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.584240] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e555e8f5-5743-48cb-9cac-9a01d99a39bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.601903] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1002.605942] env[62952]: DEBUG nova.compute.provider_tree [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.658270] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367431, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.706861] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.708090] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.708339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.709731] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba79349-f2c1-4467-b526-9560b08664ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.729220] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.758222] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.758827] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.759215] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.759572] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.759803] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.760258] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.760697] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.761060] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.761358] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.761731] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.762554] env[62952]: DEBUG nova.virt.hardware [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.774658] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Reconfiguring VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1002.776182] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78ce70da-51e1-4143-bc46-46bcd0003da0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.806209] env[62952]: DEBUG oslo_vmware.api [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1002.806209] env[62952]: value = "task-1367434" [ 1002.806209] env[62952]: _type = "Task" [ 1002.806209] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.821305] env[62952]: DEBUG oslo_vmware.api [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367434, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.911547] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1002.911547] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528aa359-c66e-3a30-2c82-e492170f392f" [ 1002.911547] env[62952]: _type = "HttpNfcLease" [ 1002.911547] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1002.912095] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1002.912095] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528aa359-c66e-3a30-2c82-e492170f392f" [ 1002.912095] env[62952]: _type = "HttpNfcLease" [ 1002.912095] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1002.914217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efeb2ce-e33d-4fde-9642-d802c25990aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.926079] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ef795-f201-3e55-2048-c5cbbb247e27/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1002.926411] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ef795-f201-3e55-2048-c5cbbb247e27/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1003.015150] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5200e5cc-b9ae-844d-717d-c01e879a84e3, 'name': SearchDatastore_Task, 'duration_secs': 0.016389} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.016357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.016697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.017040] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.017361] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.017535] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.017865] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14128be9-cb50-4765-a127-d2b911377ee0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.027801] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.029176] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.029176] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60a8c17d-1368-4ef0-bc7d-26946e25c10a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.035499] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1003.035499] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52d675fb-b322-4a7b-a49c-e32ffc3dac43" [ 1003.035499] env[62952]: _type = "Task" [ 1003.035499] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.045795] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d675fb-b322-4a7b-a49c-e32ffc3dac43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.099752] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d5549773-8ad8-40b4-be3e-d0e37c028993 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.110921] env[62952]: DEBUG nova.scheduler.client.report [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.160751] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367431, 'name': ReconfigVM_Task, 'duration_secs': 0.560099} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.161255] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 4c4845e6-79ca-4a2a-a636-842d6919e70f/4c4845e6-79ca-4a2a-a636-842d6919e70f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.162214] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2f81664-552a-4fd2-a77e-f17ba89220d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.170023] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1003.170023] env[62952]: value = "task-1367435" [ 1003.170023] env[62952]: _type = "Task" [ 1003.170023] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.182022] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367435, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.228424] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367432, 'name': ReconfigVM_Task, 'duration_secs': 0.562393} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.229600] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Reconfigured VM instance instance-00000056 to attach disk [datastore2] ba2dfb85-0ffa-4883-abf1-482441c5afb8/ba2dfb85-0ffa-4883-abf1-482441c5afb8.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.234233] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21bd3114-fa34-4692-a1e9-20edf2b9ee91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.236332] env[62952]: DEBUG nova.compute.manager [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-changed-e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.236516] env[62952]: DEBUG nova.compute.manager [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing instance network info cache due to event network-changed-e1654a86-c311-45dd-9cc6-143cf350df90. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1003.237554] env[62952]: DEBUG oslo_concurrency.lockutils [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.237932] env[62952]: DEBUG oslo_concurrency.lockutils [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.238631] env[62952]: DEBUG nova.network.neutron [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing network info cache for port e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.250827] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1003.250827] env[62952]: value = "task-1367436" [ 1003.250827] env[62952]: _type = "Task" [ 1003.250827] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.266598] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367436, 'name': Rename_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.318530] env[62952]: DEBUG oslo_vmware.api [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367434, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.329531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "1e02f16d-b86e-4533-928b-a672ce69788f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.330278] env[62952]: DEBUG oslo_concurrency.lockutils [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.330800] env[62952]: DEBUG nova.compute.manager [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.333320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d786b275-687b-4661-9f41-e1603dfcdeaa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.342484] env[62952]: DEBUG nova.compute.manager [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1003.343511] env[62952]: DEBUG nova.objects.instance [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'flavor' on Instance uuid 1e02f16d-b86e-4533-928b-a672ce69788f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.551532] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d675fb-b322-4a7b-a49c-e32ffc3dac43, 'name': SearchDatastore_Task, 'duration_secs': 0.016992} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.553046] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fd86a25-0a1a-4e6b-9a5f-c94506af366b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.560379] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1003.560379] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5250bb5f-2e72-9aba-7dce-24d46458d0ad" [ 1003.560379] env[62952]: _type = "Task" [ 1003.560379] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.571340] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5250bb5f-2e72-9aba-7dce-24d46458d0ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.610176] env[62952]: DEBUG nova.network.neutron [-] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.618283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.044s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.620794] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.974s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.621432] env[62952]: DEBUG nova.objects.instance [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lazy-loading 'resources' on Instance uuid a63ec2ac-4484-4360-962b-105a5338c1b8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.632149] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1003.656025] env[62952]: INFO nova.scheduler.client.report [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocations for instance 3bd9c356-42d6-450f-9069-b066952d9524 [ 1003.686840] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367435, 'name': Rename_Task, 'duration_secs': 0.213899} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.687230] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.690021] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a04bb27b-6282-4aab-87c5-5e6e2f83a414 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.696896] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1003.696896] env[62952]: value = "task-1367437" [ 1003.696896] env[62952]: _type = "Task" [ 1003.696896] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.708364] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.767071] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367436, 'name': Rename_Task, 'duration_secs': 0.182919} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.767715] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.769599] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a8ccfa9-09ae-43b4-9467-d57c79b69ff7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.778820] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1003.778820] env[62952]: value = "task-1367438" [ 1003.778820] env[62952]: _type = "Task" [ 1003.778820] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.795655] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.821500] env[62952]: DEBUG oslo_vmware.api [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367434, 'name': ReconfigVM_Task, 'duration_secs': 0.82252} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.822644] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.823540] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Reconfigured VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1003.850636] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.854833] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d7b3518-bf46-4f2f-9d46-fb833296b4d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.863722] env[62952]: DEBUG oslo_vmware.api [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1003.863722] env[62952]: value = "task-1367439" [ 1003.863722] env[62952]: _type = "Task" [ 1003.863722] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.878693] env[62952]: DEBUG oslo_vmware.api [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.080997] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5250bb5f-2e72-9aba-7dce-24d46458d0ad, 'name': SearchDatastore_Task, 'duration_secs': 0.01346} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.081225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.081756] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 75d640ed-c41a-4761-8867-191d8b3e1f79/75d640ed-c41a-4761-8867-191d8b3e1f79.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.083053] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f275121-0e33-48fc-a1a8-f156b4da1664 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.091250] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1004.091250] env[62952]: value = "task-1367440" [ 1004.091250] env[62952]: _type = "Task" [ 1004.091250] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.101114] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367440, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.110629] env[62952]: DEBUG nova.network.neutron [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updated VIF entry in instance network info cache for port e1654a86-c311-45dd-9cc6-143cf350df90. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.111217] env[62952]: DEBUG nova.network.neutron [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1654a86-c311-45dd-9cc6-143cf350df90", "address": "fa:16:3e:2d:97:e7", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1654a86-c3", "ovs_interfaceid": "e1654a86-c311-45dd-9cc6-143cf350df90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.113537] env[62952]: INFO nova.compute.manager [-] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Took 1.64 seconds to deallocate network for instance. [ 1004.175047] env[62952]: DEBUG oslo_concurrency.lockutils [None req-f3634b32-81a7-4ace-ad54-87a61112bfce tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "3bd9c356-42d6-450f-9069-b066952d9524" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.111s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.211032] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367437, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.291680] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367438, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.331623] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bdb18490-25b2-4d07-af2a-ffa0ad32d69d tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-ff189c9e-9e7c-4217-9c65-0f821393870a-e1654a86-c311-45dd-9cc6-143cf350df90" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.781s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.367531] env[62952]: DEBUG nova.compute.manager [req-b3ef114e-3e2d-4807-ac12-f2d9e6a42bfd req-2d98fd61-7def-4cd8-a58c-8de0acdb11fa service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Received event network-vif-plugged-3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.368198] env[62952]: DEBUG oslo_concurrency.lockutils [req-b3ef114e-3e2d-4807-ac12-f2d9e6a42bfd req-2d98fd61-7def-4cd8-a58c-8de0acdb11fa service nova] Acquiring lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.368805] env[62952]: DEBUG oslo_concurrency.lockutils [req-b3ef114e-3e2d-4807-ac12-f2d9e6a42bfd req-2d98fd61-7def-4cd8-a58c-8de0acdb11fa service nova] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.369199] env[62952]: DEBUG oslo_concurrency.lockutils [req-b3ef114e-3e2d-4807-ac12-f2d9e6a42bfd req-2d98fd61-7def-4cd8-a58c-8de0acdb11fa service nova] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.369840] env[62952]: DEBUG nova.compute.manager [req-b3ef114e-3e2d-4807-ac12-f2d9e6a42bfd req-2d98fd61-7def-4cd8-a58c-8de0acdb11fa service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] No waiting events found dispatching network-vif-plugged-3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.369933] env[62952]: WARNING nova.compute.manager [req-b3ef114e-3e2d-4807-ac12-f2d9e6a42bfd req-2d98fd61-7def-4cd8-a58c-8de0acdb11fa service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Received unexpected event network-vif-plugged-3f88a864-36dc-4fc8-8073-977444f51ed8 for instance with vm_state building and task_state spawning. [ 1004.394556] env[62952]: DEBUG oslo_vmware.api [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367439, 'name': PowerOffVM_Task, 'duration_secs': 0.328258} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.399786] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.400283] env[62952]: DEBUG nova.compute.manager [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.402510] env[62952]: DEBUG nova.network.neutron [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Successfully updated port: 3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.407505] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0464c72-721f-437f-8511-7cfbc7d17cf4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.485070] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ec0a2a-5be3-404c-96a2-bf395618eced {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.500299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bf5c44-9400-4e75-b24d-cd7ebd61dfae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.548544] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13673e1-01da-4d2a-b447-196aba25a34d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.559102] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd16257d-b058-4b49-a869-47fa05753de9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.577526] env[62952]: DEBUG nova.compute.provider_tree [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.606836] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367440, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.614986] env[62952]: DEBUG oslo_concurrency.lockutils [req-0106b647-53fb-4e9c-bbd5-f8df05557a11 req-8bcd18df-a965-4900-aa4b-75f62fb581e6 service nova] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.622591] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.707416] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367437, 'name': PowerOnVM_Task, 'duration_secs': 0.847716} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.707744] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.707993] env[62952]: INFO nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Took 11.35 seconds to spawn the instance on the hypervisor. [ 1004.708216] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.709081] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e05a9e4-07dd-4d51-a71c-4892c96439a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.792173] env[62952]: DEBUG oslo_vmware.api [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367438, 'name': PowerOnVM_Task, 'duration_secs': 0.685249} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.792495] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.792713] env[62952]: INFO nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Took 9.09 seconds to spawn the instance on the hypervisor. [ 1004.792974] env[62952]: DEBUG nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.793827] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba781a8-7363-4c2a-8d5a-b7a5edd1e063 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.916752] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.917700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.917814] env[62952]: DEBUG nova.network.neutron [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1004.931281] env[62952]: DEBUG oslo_concurrency.lockutils [None req-56551861-3b96-43e0-9a59-20de0d71ba00 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.601s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.081755] env[62952]: DEBUG nova.scheduler.client.report [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.102039] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367440, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620829} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.102250] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 75d640ed-c41a-4761-8867-191d8b3e1f79/75d640ed-c41a-4761-8867-191d8b3e1f79.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.102364] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.102621] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3ab216b-091e-4593-9846-e14dd287d822 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.110646] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1005.110646] env[62952]: value = "task-1367441" [ 1005.110646] env[62952]: _type = "Task" [ 1005.110646] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.120209] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367441, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.228507] env[62952]: INFO nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Took 37.52 seconds to build instance. [ 1005.312140] env[62952]: INFO nova.compute.manager [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Took 37.54 seconds to build instance. [ 1005.465259] env[62952]: DEBUG nova.network.neutron [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1005.587949] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.967s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.590657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.577s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.590854] env[62952]: DEBUG nova.objects.instance [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1005.623779] env[62952]: INFO nova.scheduler.client.report [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Deleted allocations for instance a63ec2ac-4484-4360-962b-105a5338c1b8 [ 1005.625320] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367441, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113688} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.625611] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.626458] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bef25d9-a8ce-4dac-b1a2-627f777fd42e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.650861] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 75d640ed-c41a-4761-8867-191d8b3e1f79/75d640ed-c41a-4761-8867-191d8b3e1f79.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.651750] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61887db0-856e-404a-b4ca-4ea7f3dc1289 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.672877] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1005.672877] env[62952]: value = "task-1367442" [ 1005.672877] env[62952]: _type = "Task" [ 1005.672877] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.682010] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367442, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.692810] env[62952]: DEBUG nova.network.neutron [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance_info_cache with network_info: [{"id": "3f88a864-36dc-4fc8-8073-977444f51ed8", "address": "fa:16:3e:70:82:3d", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f88a864-36", "ovs_interfaceid": "3f88a864-36dc-4fc8-8073-977444f51ed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.730717] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.029s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.772452] env[62952]: DEBUG nova.compute.manager [req-5f2a0907-8827-4f90-8ba5-5eeb7a2c90de req-400f3c6b-c17b-4650-9713-8c9a3aa542a2 service nova] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Received event network-vif-deleted-3582454d-1ad8-4562-8f55-42a3f083d328 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.813252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef020cd5-19e3-4748-a29e-a1557b0525dd tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.051s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.154969] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b90465a1-447c-4c64-b15a-ab3d71c4d91d tempest-ListServerFiltersTestJSON-200103318 tempest-ListServerFiltersTestJSON-200103318-project-member] Lock "a63ec2ac-4484-4360-962b-105a5338c1b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.986s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.184544] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367442, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.196100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.196742] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Instance network_info: |[{"id": "3f88a864-36dc-4fc8-8073-977444f51ed8", "address": "fa:16:3e:70:82:3d", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f88a864-36", "ovs_interfaceid": "3f88a864-36dc-4fc8-8073-977444f51ed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1006.600100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-65bbfbe7-03c7-4562-b787-802ce149f75f tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.601049] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.892s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.601354] env[62952]: DEBUG nova.objects.instance [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lazy-loading 'resources' on Instance uuid 97995e38-b27e-478a-8553-eb1c844bb0a6 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.685386] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367442, 'name': ReconfigVM_Task, 'duration_secs': 0.537984} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.685767] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 75d640ed-c41a-4761-8867-191d8b3e1f79/75d640ed-c41a-4761-8867-191d8b3e1f79.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.686506] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-409e2870-991b-4959-aa86-92b4ef6683df {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.694268] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1006.694268] env[62952]: value = "task-1367443" [ 1006.694268] env[62952]: _type = "Task" [ 1006.694268] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.702965] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367443, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.730868] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "interface-ff189c9e-9e7c-4217-9c65-0f821393870a-e1654a86-c311-45dd-9cc6-143cf350df90" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.731265] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-ff189c9e-9e7c-4217-9c65-0f821393870a-e1654a86-c311-45dd-9cc6-143cf350df90" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.208277] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367443, 'name': Rename_Task, 'duration_secs': 0.198094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.208632] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.208918] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66fe6fad-f696-493d-98ac-6676e11ce8b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.216513] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1007.216513] env[62952]: value = "task-1367444" [ 1007.216513] env[62952]: _type = "Task" [ 1007.216513] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.228496] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.235995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.236252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.237488] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b5df3a-28a1-4fea-a799-83a520da19d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.260639] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824cb903-ad22-432e-ad45-e5a4ce2810fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.290760] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Reconfiguring VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1007.294634] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f63b042-6aa4-4311-ae3e-bfc90de8ef91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.317802] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1007.317802] env[62952]: value = "task-1367445" [ 1007.317802] env[62952]: _type = "Task" [ 1007.317802] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.332134] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.414293] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a670e4c-9921-45a2-adc1-c9567dc92319 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.421879] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f93fb50-76e2-468b-ac94-aa10eb185577 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.455112] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87234be2-9054-47a9-84a3-d05938946961 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.462966] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461894b9-ee59-4e09-b2ae-b562cbfd607e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.478193] env[62952]: DEBUG nova.compute.provider_tree [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.726781] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367444, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.829228] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.981818] env[62952]: DEBUG nova.scheduler.client.report [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.227017] env[62952]: DEBUG oslo_vmware.api [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367444, 'name': PowerOnVM_Task, 'duration_secs': 0.758272} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.227402] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1008.227519] env[62952]: INFO nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Took 10.13 seconds to spawn the instance on the hypervisor. [ 1008.227669] env[62952]: DEBUG nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.228509] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a77b356-8cae-4f97-af3a-e8c03361288c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.333206] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.487319] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.489833] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.374s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.491484] env[62952]: INFO nova.compute.claims [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.514110] env[62952]: INFO nova.scheduler.client.report [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleted allocations for instance 97995e38-b27e-478a-8553-eb1c844bb0a6 [ 1008.746024] env[62952]: INFO nova.compute.manager [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Took 39.16 seconds to build instance. [ 1008.832335] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.021806] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e2010-616b-4b5d-b6f5-8efdb0c18eae tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "97995e38-b27e-478a-8553-eb1c844bb0a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.049s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.248907] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1413a3c4-e5e9-48bd-a6d2-1d5c94fe3451 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.318s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.330888] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.689382] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b48be0a-13a0-4b41-a743-05c88c0b738d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.697015] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8724af2d-b183-4738-ab67-9e22142aac77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.727855] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfc2cc2-c089-4191-a535-2aa2bb417b8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.735644] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b262f9df-6c0c-472c-98fe-647acde19fe1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.749183] env[62952]: DEBUG nova.compute.provider_tree [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.831530] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.252982] env[62952]: DEBUG nova.scheduler.client.report [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.331610] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.758544] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.758779] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1010.763039] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.140s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.763039] env[62952]: DEBUG nova.objects.instance [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid da749732-72d8-4180-aedf-73a17fa9dea6 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.801188] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.801605] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.801817] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.801999] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.802188] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.802421] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.802657] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.802852] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.803073] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.803260] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.803463] env[62952]: DEBUG nova.virt.hardware [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.806756] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8be86f6-6937-49ce-99df-882ed8194001 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.819655] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ef795-f201-3e55-2048-c5cbbb247e27/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1010.822981] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4e2491-6782-4648-83b0-299a0df6ed77 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.827138] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b626f5-dc9e-4e41-8787-9989c3fb5e52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.839057] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ef795-f201-3e55-2048-c5cbbb247e27/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1010.839247] env[62952]: ERROR oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ef795-f201-3e55-2048-c5cbbb247e27/disk-0.vmdk due to incomplete transfer. [ 1010.850336] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2a514a3d-7c92-4a9c-a476-ca9ad980931d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.852185] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.853066] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:82:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f88a864-36dc-4fc8-8073-977444f51ed8', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.861018] env[62952]: DEBUG oslo.service.loopingcall [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.861359] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1010.862625] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d08b541-4aa3-479e-834a-a9391c8f9343 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.881035] env[62952]: DEBUG oslo_vmware.rw_handles [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ef795-f201-3e55-2048-c5cbbb247e27/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1010.881280] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Uploaded image 1de1bca9-aeab-4b87-85e8-152a9124e537 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1010.883211] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1010.887844] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a84ed5d1-1a06-4625-94af-efaee51f93c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.893015] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.893015] env[62952]: value = "task-1367446" [ 1010.893015] env[62952]: _type = "Task" [ 1010.893015] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.904769] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1010.904769] env[62952]: value = "task-1367447" [ 1010.904769] env[62952]: _type = "Task" [ 1010.904769] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.916237] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367446, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.922720] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367447, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.927938] env[62952]: DEBUG nova.compute.manager [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Received event network-changed-3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.928273] env[62952]: DEBUG nova.compute.manager [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Refreshing instance network info cache due to event network-changed-3f88a864-36dc-4fc8-8073-977444f51ed8. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1010.928494] env[62952]: DEBUG oslo_concurrency.lockutils [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] Acquiring lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.928590] env[62952]: DEBUG oslo_concurrency.lockutils [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] Acquired lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.928941] env[62952]: DEBUG nova.network.neutron [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Refreshing network info cache for port 3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.071583] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f10991-5525-8b0e-4d9c-71d8c9eb6126/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1011.073791] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cbbb6f-48d4-4429-b4f5-f2f25e1ab9d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.082066] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f10991-5525-8b0e-4d9c-71d8c9eb6126/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1011.082066] env[62952]: ERROR oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f10991-5525-8b0e-4d9c-71d8c9eb6126/disk-0.vmdk due to incomplete transfer. [ 1011.082066] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6656524c-ceb6-447a-b8c5-56dc7cf3bacf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.091280] env[62952]: DEBUG oslo_vmware.rw_handles [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f10991-5525-8b0e-4d9c-71d8c9eb6126/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1011.091280] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Uploaded image 497ca4b0-de44-46b4-b59e-8ecf1e60917f to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1011.091608] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1011.091997] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dbca2dab-e3fc-4dc2-8e95-573b7c8e9ed8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.102907] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1011.102907] env[62952]: value = "task-1367448" [ 1011.102907] env[62952]: _type = "Task" [ 1011.102907] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.111482] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367448, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.266438] env[62952]: DEBUG nova.compute.utils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.274554] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.274758] env[62952]: DEBUG nova.network.neutron [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.326158] env[62952]: DEBUG nova.policy [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f24a30e1160a43d5be6cde39b8f79edd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '094ca56818604bee9c1f75d1493ab3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.341092] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.388814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "4c4845e6-79ca-4a2a-a636-842d6919e70f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.388814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.388814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "4c4845e6-79ca-4a2a-a636-842d6919e70f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.388814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.389090] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.395921] env[62952]: INFO nova.compute.manager [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Terminating instance [ 1011.403245] env[62952]: DEBUG nova.compute.manager [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1011.403529] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.404898] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf0cdc5-417b-463b-8dd7-1d71bc807973 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.409330] env[62952]: INFO nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Rebuilding instance [ 1011.423566] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367446, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.437851] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367447, 'name': Destroy_Task, 'duration_secs': 0.366992} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.438206] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.441538] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Destroyed the VM [ 1011.442863] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1011.442863] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a45545e-7c4d-41f5-ba5a-d082979379e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.444752] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-471a3d29-e8f7-4a8a-af50-b863fd9e3410 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.457335] env[62952]: DEBUG oslo_vmware.api [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1011.457335] env[62952]: value = "task-1367449" [ 1011.457335] env[62952]: _type = "Task" [ 1011.457335] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.459477] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1011.459477] env[62952]: value = "task-1367450" [ 1011.459477] env[62952]: _type = "Task" [ 1011.459477] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.478494] env[62952]: DEBUG oslo_vmware.api [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.485045] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367450, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.485618] env[62952]: DEBUG nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.486713] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da42ce88-a415-4c21-bec9-5a4ad9ca35d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.586176] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938db93a-53dc-4b49-a1eb-4c00362c62e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.594574] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a2540a-027f-48a8-8ba1-d6801bab04b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.631776] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a434e1e6-a28f-48da-8c80-67925a94fcdd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.640190] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367448, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.643864] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b27063-ae0a-4ab2-a27e-ba4b90986252 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.660083] env[62952]: DEBUG nova.compute.provider_tree [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.697582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.697929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.698103] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.698325] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.698506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.700655] env[62952]: INFO nova.compute.manager [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Terminating instance [ 1011.704874] env[62952]: DEBUG nova.compute.manager [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1011.705090] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.705980] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2129a6-c239-4580-9a35-84e8af1ecc5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.714247] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.714558] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fab142b9-f495-4459-8a2b-95ff389fa71e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.721332] env[62952]: DEBUG oslo_vmware.api [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1011.721332] env[62952]: value = "task-1367451" [ 1011.721332] env[62952]: _type = "Task" [ 1011.721332] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.731319] env[62952]: DEBUG oslo_vmware.api [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.775070] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.813746] env[62952]: DEBUG nova.network.neutron [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updated VIF entry in instance network info cache for port 3f88a864-36dc-4fc8-8073-977444f51ed8. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.814183] env[62952]: DEBUG nova.network.neutron [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance_info_cache with network_info: [{"id": "3f88a864-36dc-4fc8-8073-977444f51ed8", "address": "fa:16:3e:70:82:3d", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f88a864-36", "ovs_interfaceid": "3f88a864-36dc-4fc8-8073-977444f51ed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.840280] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.869506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "82401449-347e-4917-a307-d5d7bc048f7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.869823] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "82401449-347e-4917-a307-d5d7bc048f7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.870102] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "82401449-347e-4917-a307-d5d7bc048f7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.870316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "82401449-347e-4917-a307-d5d7bc048f7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.870527] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "82401449-347e-4917-a307-d5d7bc048f7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.874304] env[62952]: INFO nova.compute.manager [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Terminating instance [ 1011.876515] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "refresh_cache-82401449-347e-4917-a307-d5d7bc048f7c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.876727] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquired lock "refresh_cache-82401449-347e-4917-a307-d5d7bc048f7c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.876968] env[62952]: DEBUG nova.network.neutron [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.907025] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367446, 'name': CreateVM_Task, 'duration_secs': 0.634869} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.907199] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.908790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.908790] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.908989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.909467] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d4dd118-a50c-4eb6-888b-2c016c585aa8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.914232] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1011.914232] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524cbffa-a1fa-2416-35f1-933730e4b246" [ 1011.914232] env[62952]: _type = "Task" [ 1011.914232] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.926435] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524cbffa-a1fa-2416-35f1-933730e4b246, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.973879] env[62952]: DEBUG oslo_vmware.api [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367449, 'name': PowerOffVM_Task, 'duration_secs': 0.242936} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.974157] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.974361] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1011.977354] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5dbda88c-4cf9-44fd-898e-57c2e1996fe4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.984895] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367450, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.002655] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.003155] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-885cba59-0f46-4136-9409-0917405aa89f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.011062] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1012.011062] env[62952]: value = "task-1367453" [ 1012.011062] env[62952]: _type = "Task" [ 1012.011062] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.021893] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1012.022295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.023098] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68aac70-f767-45f3-915f-b3c32ef59c23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.030564] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.030834] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5eb1261-904f-4352-a405-6f1bb86f3763 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.043453] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.043859] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.043859] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleting the datastore file [datastore1] 4c4845e6-79ca-4a2a-a636-842d6919e70f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.044196] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-710ec2f5-a245-4b0c-a808-8e9cf826d44b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.051552] env[62952]: DEBUG oslo_vmware.api [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1012.051552] env[62952]: value = "task-1367455" [ 1012.051552] env[62952]: _type = "Task" [ 1012.051552] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.052439] env[62952]: DEBUG nova.network.neutron [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Successfully created port: 143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.062674] env[62952]: DEBUG oslo_vmware.api [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.122026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.122026] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.122026] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.122026] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bed1c7c-06e0-435a-8d90-a3f4bb1aadc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.129568] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1012.129568] env[62952]: value = "task-1367456" [ 1012.129568] env[62952]: _type = "Task" [ 1012.129568] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.151699] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367448, 'name': Destroy_Task, 'duration_secs': 0.780533} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.152895] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367456, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.152895] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroyed the VM [ 1012.153308] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1012.160020] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2eff8095-6f59-4475-9288-1d3fb58c7683 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.167729] env[62952]: DEBUG nova.scheduler.client.report [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.173383] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1012.173383] env[62952]: value = "task-1367457" [ 1012.173383] env[62952]: _type = "Task" [ 1012.173383] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.184612] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367457, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.231721] env[62952]: DEBUG oslo_vmware.api [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367451, 'name': PowerOffVM_Task, 'duration_secs': 0.297749} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.232083] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.232333] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.232671] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9dc9e9d2-9772-42f3-b972-223be7083d21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.304793] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.305210] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.305306] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleting the datastore file [datastore2] ba2dfb85-0ffa-4883-abf1-482441c5afb8 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.305504] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01e1ade0-ec3e-434e-acea-c08e48a49d30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.312326] env[62952]: DEBUG oslo_vmware.api [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for the task: (returnval){ [ 1012.312326] env[62952]: value = "task-1367459" [ 1012.312326] env[62952]: _type = "Task" [ 1012.312326] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.318162] env[62952]: DEBUG oslo_concurrency.lockutils [req-d4c5dc62-0bae-44e5-ae97-8487f35a8e4d req-fd19edeb-f95c-4e57-8aec-e71186bf2327 service nova] Releasing lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.321598] env[62952]: DEBUG oslo_vmware.api [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.336976] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.401894] env[62952]: DEBUG nova.network.neutron [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1012.428104] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524cbffa-a1fa-2416-35f1-933730e4b246, 'name': SearchDatastore_Task, 'duration_secs': 0.010145} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.428568] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.429319] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.429631] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.429797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.429981] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.430286] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21732eae-7d79-4ebe-979d-38ee247c96d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.439289] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.439516] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.440265] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21ff11e7-dac5-4aba-abd9-515bc5e5d9ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.446226] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1012.446226] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a62ca2-067e-6467-a74a-ac40a8ec9cea" [ 1012.446226] env[62952]: _type = "Task" [ 1012.446226] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.453253] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a62ca2-067e-6467-a74a-ac40a8ec9cea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.465265] env[62952]: DEBUG nova.network.neutron [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.474848] env[62952]: DEBUG oslo_vmware.api [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367450, 'name': RemoveSnapshot_Task, 'duration_secs': 0.729538} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.475940] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1012.475940] env[62952]: INFO nova.compute.manager [None req-2f043fb6-8bd7-4813-ad2a-1685e44bf290 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Took 15.83 seconds to snapshot the instance on the hypervisor. [ 1012.563509] env[62952]: DEBUG oslo_vmware.api [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199776} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.563765] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.563953] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.564157] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.564342] env[62952]: INFO nova.compute.manager [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1012.564584] env[62952]: DEBUG oslo.service.loopingcall [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.564775] env[62952]: DEBUG nova.compute.manager [-] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1012.564868] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1012.646731] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367456, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170228} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.648033] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.648033] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.648033] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.675625] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.693962] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367457, 'name': RemoveSnapshot_Task} progress is 31%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.699822] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.700157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.708775] env[62952]: INFO nova.scheduler.client.report [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance da749732-72d8-4180-aedf-73a17fa9dea6 [ 1012.785561] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1012.816022] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.816202] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.816420] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.816699] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.816880] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.817066] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.817303] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.817481] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.817668] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.817844] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.818550] env[62952]: DEBUG nova.virt.hardware [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.819059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ff3568-6c9d-42ff-be60-7c02031fdef4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.838113] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f19290e-1159-4d06-b3cd-c5a1d62c413b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.841625] env[62952]: DEBUG oslo_vmware.api [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Task: {'id': task-1367459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158001} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.841924] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.842134] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.842336] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.842518] env[62952]: INFO nova.compute.manager [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1012.842783] env[62952]: DEBUG oslo.service.loopingcall [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.843580] env[62952]: DEBUG nova.compute.manager [-] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1012.843677] env[62952]: DEBUG nova.network.neutron [-] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1012.855867] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.929645] env[62952]: DEBUG nova.compute.manager [req-e6a3f69b-e009-42cc-a28c-22e87c26d895 req-22fc5d17-df50-4f63-8a0f-f19e65aeab75 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Received event network-vif-deleted-567af111-f74a-4736-b704-de3514ef2899 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.930294] env[62952]: INFO nova.compute.manager [req-e6a3f69b-e009-42cc-a28c-22e87c26d895 req-22fc5d17-df50-4f63-8a0f-f19e65aeab75 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Neutron deleted interface 567af111-f74a-4736-b704-de3514ef2899; detaching it from the instance and deleting it from the info cache [ 1012.930498] env[62952]: DEBUG nova.network.neutron [req-e6a3f69b-e009-42cc-a28c-22e87c26d895 req-22fc5d17-df50-4f63-8a0f-f19e65aeab75 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.957449] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a62ca2-067e-6467-a74a-ac40a8ec9cea, 'name': SearchDatastore_Task, 'duration_secs': 0.008305} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.958589] env[62952]: DEBUG nova.compute.manager [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Received event network-changed-6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.958796] env[62952]: DEBUG nova.compute.manager [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Refreshing instance network info cache due to event network-changed-6e4543b9-4601-4109-b4fc-14eee838c035. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.959037] env[62952]: DEBUG oslo_concurrency.lockutils [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] Acquiring lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.959184] env[62952]: DEBUG oslo_concurrency.lockutils [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] Acquired lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.960097] env[62952]: DEBUG nova.network.neutron [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Refreshing network info cache for port 6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1012.961025] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bead08b9-f034-4b7c-abea-ad923abb8ee3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.966863] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1012.966863] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52352b23-1b25-a440-ca13-387b8cb9513d" [ 1012.966863] env[62952]: _type = "Task" [ 1012.966863] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.974120] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Releasing lock "refresh_cache-82401449-347e-4917-a307-d5d7bc048f7c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.974889] env[62952]: DEBUG nova.compute.manager [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1012.975127] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.976526] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea047ea3-6ac3-4a70-8404-861f80e10ee9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.987350] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52352b23-1b25-a440-ca13-387b8cb9513d, 'name': SearchDatastore_Task, 'duration_secs': 0.010845} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.988451] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.988705] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4/e0384d0d-6738-47f7-b14d-994cae2d1fe4.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1012.991045] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e399519-d8ce-4673-9ea1-458748542651 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.993035] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.993678] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59d482e9-adda-4d0d-912b-423085ca60cb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.999326] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1012.999326] env[62952]: value = "task-1367460" [ 1012.999326] env[62952]: _type = "Task" [ 1012.999326] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.000644] env[62952]: DEBUG oslo_vmware.api [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 1013.000644] env[62952]: value = "task-1367461" [ 1013.000644] env[62952]: _type = "Task" [ 1013.000644] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.013055] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367460, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.016065] env[62952]: DEBUG oslo_vmware.api [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367461, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.191821] env[62952]: DEBUG oslo_vmware.api [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367457, 'name': RemoveSnapshot_Task, 'duration_secs': 1.01023} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.193286] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1013.193649] env[62952]: INFO nova.compute.manager [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 16.46 seconds to snapshot the instance on the hypervisor. [ 1013.203321] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1013.217200] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ccc5f349-c620-47b1-9361-5ea6db1bf02f tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "da749732-72d8-4180-aedf-73a17fa9dea6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.942s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.340412] env[62952]: DEBUG oslo_vmware.api [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367445, 'name': ReconfigVM_Task, 'duration_secs': 5.869372} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.340897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.341232] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Reconfigured VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1013.384432] env[62952]: DEBUG nova.network.neutron [-] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.434037] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-538be98a-d20c-43e8-888c-a37f7b12ed3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.442987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adb5d0c-f8b8-44d8-a258-160212118c1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.476891] env[62952]: DEBUG nova.compute.manager [req-e6a3f69b-e009-42cc-a28c-22e87c26d895 req-22fc5d17-df50-4f63-8a0f-f19e65aeab75 service nova] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Detach interface failed, port_id=567af111-f74a-4736-b704-de3514ef2899, reason: Instance 4c4845e6-79ca-4a2a-a636-842d6919e70f could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1013.519679] env[62952]: DEBUG oslo_vmware.api [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367461, 'name': PowerOffVM_Task, 'duration_secs': 0.265374} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.519945] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367460, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51446} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.520422] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.520608] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.520876] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4/e0384d0d-6738-47f7-b14d-994cae2d1fe4.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.521094] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.521343] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f61ddd9-66e1-4074-8670-2118484dd255 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.522860] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-910c2111-4bd6-48dd-8188-5eaf66250cf7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.530558] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1013.530558] env[62952]: value = "task-1367463" [ 1013.530558] env[62952]: _type = "Task" [ 1013.530558] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.547330] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.547832] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.547832] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleting the datastore file [datastore1] 82401449-347e-4917-a307-d5d7bc048f7c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.552697] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-398e0e19-21f4-4601-ac79-9db9d6e7dc10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.553519] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.557975] env[62952]: DEBUG oslo_vmware.api [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for the task: (returnval){ [ 1013.557975] env[62952]: value = "task-1367464" [ 1013.557975] env[62952]: _type = "Task" [ 1013.557975] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.567101] env[62952]: DEBUG oslo_vmware.api [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.591690] env[62952]: DEBUG nova.network.neutron [-] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.681079] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.681338] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.681501] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.681682] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.681829] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.681975] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.682445] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.683058] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.683058] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.683058] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.683272] env[62952]: DEBUG nova.virt.hardware [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.684432] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4035fbe5-3d63-4668-9a7d-f2604a8b5012 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.692430] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2a06a4-1692-4b23-9d46-f165c3041912 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.712556] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:e3:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97ceb54a-06d4-4154-9965-95093341cdb4', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.719575] env[62952]: DEBUG oslo.service.loopingcall [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.725787] env[62952]: DEBUG nova.network.neutron [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Successfully updated port: 143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.729370] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.732345] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a69f9214-9a6b-49fd-a9ae-2a1de3ef0917 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.755649] env[62952]: DEBUG nova.compute.manager [req-3e1c02e1-361b-4024-b55c-e8fd71524268 req-07946cf0-19c6-4871-918c-e110e438907a service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Received event network-vif-plugged-143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.757502] env[62952]: DEBUG oslo_concurrency.lockutils [req-3e1c02e1-361b-4024-b55c-e8fd71524268 req-07946cf0-19c6-4871-918c-e110e438907a service nova] Acquiring lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.757502] env[62952]: DEBUG oslo_concurrency.lockutils [req-3e1c02e1-361b-4024-b55c-e8fd71524268 req-07946cf0-19c6-4871-918c-e110e438907a service nova] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.757502] env[62952]: DEBUG oslo_concurrency.lockutils [req-3e1c02e1-361b-4024-b55c-e8fd71524268 req-07946cf0-19c6-4871-918c-e110e438907a service nova] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.757502] env[62952]: DEBUG nova.compute.manager [req-3e1c02e1-361b-4024-b55c-e8fd71524268 req-07946cf0-19c6-4871-918c-e110e438907a service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] No waiting events found dispatching network-vif-plugged-143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1013.757502] env[62952]: WARNING nova.compute.manager [req-3e1c02e1-361b-4024-b55c-e8fd71524268 req-07946cf0-19c6-4871-918c-e110e438907a service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Received unexpected event network-vif-plugged-143465ea-10eb-4235-a204-f1a1003a57a6 for instance with vm_state building and task_state spawning. [ 1013.761797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.761945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.764068] env[62952]: INFO nova.compute.claims [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.771707] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.771707] env[62952]: value = "task-1367465" [ 1013.771707] env[62952]: _type = "Task" [ 1013.771707] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.784832] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367465, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.786671] env[62952]: DEBUG nova.compute.manager [None req-347342f4-aab6-403f-aa8c-f58577386e6c tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Found 1 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1013.888023] env[62952]: INFO nova.compute.manager [-] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Took 1.32 seconds to deallocate network for instance. [ 1014.011789] env[62952]: DEBUG nova.network.neutron [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updated VIF entry in instance network info cache for port 6e4543b9-4601-4109-b4fc-14eee838c035. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.012189] env[62952]: DEBUG nova.network.neutron [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updating instance_info_cache with network_info: [{"id": "6e4543b9-4601-4109-b4fc-14eee838c035", "address": "fa:16:3e:89:6a:b0", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e4543b9-46", "ovs_interfaceid": "6e4543b9-4601-4109-b4fc-14eee838c035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.040270] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06962} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.040531] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.041344] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bed06b-6ad6-40ba-ac72-04df7f1c68ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.063084] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4/e0384d0d-6738-47f7-b14d-994cae2d1fe4.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.063255] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bf90f41-1569-49eb-b777-c563ec818b6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.088310] env[62952]: DEBUG oslo_vmware.api [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Task: {'id': task-1367464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098659} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.088655] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.088869] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.089068] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.089252] env[62952]: INFO nova.compute.manager [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1014.089504] env[62952]: DEBUG oslo.service.loopingcall [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.089707] env[62952]: DEBUG nova.compute.manager [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.089802] env[62952]: DEBUG nova.network.neutron [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.093383] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1014.093383] env[62952]: value = "task-1367466" [ 1014.093383] env[62952]: _type = "Task" [ 1014.093383] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.093830] env[62952]: INFO nova.compute.manager [-] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Took 1.25 seconds to deallocate network for instance. [ 1014.106649] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367466, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.107413] env[62952]: DEBUG nova.network.neutron [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.231932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-1df00ddb-64fd-4ad3-be02-072a5aaad6f2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.232118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-1df00ddb-64fd-4ad3-be02-072a5aaad6f2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.232295] env[62952]: DEBUG nova.network.neutron [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.284023] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367465, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.412280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.427477] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.427766] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.514940] env[62952]: DEBUG oslo_concurrency.lockutils [req-f216d0f7-608e-4bcf-bc0c-2b803ec3fac9 req-b4422d8e-2d2a-4bb6-b29f-d6fc3a13d8bc service nova] Releasing lock "refresh_cache-75d640ed-c41a-4761-8867-191d8b3e1f79" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.604652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.604978] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367466, 'name': ReconfigVM_Task, 'duration_secs': 0.386666} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.605228] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Reconfigured VM instance instance-00000058 to attach disk [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4/e0384d0d-6738-47f7-b14d-994cae2d1fe4.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.605830] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6a3267a-4a9c-4b5e-a209-d6ae4b38e559 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.610036] env[62952]: DEBUG nova.network.neutron [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.614270] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1014.614270] env[62952]: value = "task-1367467" [ 1014.614270] env[62952]: _type = "Task" [ 1014.614270] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.624403] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367467, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.786820] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367465, 'name': CreateVM_Task, 'duration_secs': 0.685625} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.787758] env[62952]: DEBUG nova.network.neutron [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.789661] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.790438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.790638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.790989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.791599] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e75763ab-f151-4485-806a-446d0f6a9012 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.800442] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1014.800442] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]527503c3-6e27-9448-8d5b-0132ee299d65" [ 1014.800442] env[62952]: _type = "Task" [ 1014.800442] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.813131] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527503c3-6e27-9448-8d5b-0132ee299d65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.932942] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1014.985588] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.989019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.989019] env[62952]: DEBUG nova.network.neutron [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.994327] env[62952]: DEBUG nova.network.neutron [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Updating instance_info_cache with network_info: [{"id": "143465ea-10eb-4235-a204-f1a1003a57a6", "address": "fa:16:3e:90:32:4a", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap143465ea-10", "ovs_interfaceid": "143465ea-10eb-4235-a204-f1a1003a57a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.998046] env[62952]: DEBUG nova.compute.manager [req-f155ab72-9e9e-4287-93e1-84b71c2406bd req-81ca5a51-4472-4cdb-8c62-fc1d69917102 service nova] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Received event network-vif-deleted-cfd7dd45-fd8c-45d4-a276-ed89144b694f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.015026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d83049-c0ad-47c7-a19c-0c050ba53442 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.023926] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed06da25-2d09-4f51-abc9-9f34040002c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.055927] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0162c190-215e-49cb-9b78-c1bdeab162b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.064347] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12a9a6d-42a3-427f-911c-0469d48b758c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.078327] env[62952]: DEBUG nova.compute.provider_tree [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.112529] env[62952]: INFO nova.compute.manager [-] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Took 1.02 seconds to deallocate network for instance. [ 1015.126183] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367467, 'name': Rename_Task, 'duration_secs': 0.164569} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.126456] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.126680] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23aed0b5-e865-4bcb-bc33-fc2481e7154f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.135099] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1015.135099] env[62952]: value = "task-1367468" [ 1015.135099] env[62952]: _type = "Task" [ 1015.135099] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.143185] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.155787] env[62952]: DEBUG nova.compute.manager [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.156642] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee89de65-7aae-4a8b-a1da-60349e21a1a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.168352] env[62952]: DEBUG nova.compute.manager [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Received event network-changed-143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.168582] env[62952]: DEBUG nova.compute.manager [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Refreshing instance network info cache due to event network-changed-143465ea-10eb-4235-a204-f1a1003a57a6. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.168947] env[62952]: DEBUG oslo_concurrency.lockutils [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] Acquiring lock "refresh_cache-1df00ddb-64fd-4ad3-be02-072a5aaad6f2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.314307] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527503c3-6e27-9448-8d5b-0132ee299d65, 'name': SearchDatastore_Task, 'duration_secs': 0.016308} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.314624] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.314864] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.315119] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.315276] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.315525] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.315757] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28329e78-49ea-4270-8a0f-f4eac7f095f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.324456] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.324644] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.325364] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64ef3ae9-a471-473c-a188-934600fa0c91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.331838] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1015.331838] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b59d30-3786-395f-a813-0e299dd6ce9c" [ 1015.331838] env[62952]: _type = "Task" [ 1015.331838] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.339693] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b59d30-3786-395f-a813-0e299dd6ce9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.451921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.500616] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-1df00ddb-64fd-4ad3-be02-072a5aaad6f2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.500616] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Instance network_info: |[{"id": "143465ea-10eb-4235-a204-f1a1003a57a6", "address": "fa:16:3e:90:32:4a", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap143465ea-10", "ovs_interfaceid": "143465ea-10eb-4235-a204-f1a1003a57a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.501012] env[62952]: DEBUG oslo_concurrency.lockutils [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] Acquired lock "refresh_cache-1df00ddb-64fd-4ad3-be02-072a5aaad6f2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.501012] env[62952]: DEBUG nova.network.neutron [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Refreshing network info cache for port 143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.506074] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:32:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97113f46-d648-4613-b233-069acba18198', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '143465ea-10eb-4235-a204-f1a1003a57a6', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.514505] env[62952]: DEBUG oslo.service.loopingcall [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.517839] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.518357] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a2388c5-684a-45ff-baaf-9362fb8c5d2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.544494] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.544494] env[62952]: value = "task-1367469" [ 1015.544494] env[62952]: _type = "Task" [ 1015.544494] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.553204] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367469, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.583364] env[62952]: DEBUG nova.scheduler.client.report [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.621633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.648746] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367468, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.670809] env[62952]: INFO nova.compute.manager [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] instance snapshotting [ 1015.670809] env[62952]: DEBUG nova.objects.instance [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'flavor' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.842899] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b59d30-3786-395f-a813-0e299dd6ce9c, 'name': SearchDatastore_Task, 'duration_secs': 0.010396} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.843762] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9171ae44-f146-48da-b763-4ae0a2f56f85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.849995] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1015.849995] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52fb63d7-ca9b-6a90-a860-8877f94a2caa" [ 1015.849995] env[62952]: _type = "Task" [ 1015.849995] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.858794] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fb63d7-ca9b-6a90-a860-8877f94a2caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.859711] env[62952]: INFO nova.network.neutron [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Port e1654a86-c311-45dd-9cc6-143cf350df90 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1015.860079] env[62952]: DEBUG nova.network.neutron [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.907054] env[62952]: DEBUG nova.network.neutron [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Updated VIF entry in instance network info cache for port 143465ea-10eb-4235-a204-f1a1003a57a6. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1015.907446] env[62952]: DEBUG nova.network.neutron [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Updating instance_info_cache with network_info: [{"id": "143465ea-10eb-4235-a204-f1a1003a57a6", "address": "fa:16:3e:90:32:4a", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap143465ea-10", "ovs_interfaceid": "143465ea-10eb-4235-a204-f1a1003a57a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.058373] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367469, 'name': CreateVM_Task, 'duration_secs': 0.365335} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.058599] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1016.059347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.059576] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.059942] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1016.060215] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce6e221d-9af6-4baa-bfb7-7189697cda61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.069065] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1016.069065] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52f54915-ccdc-37dd-5a26-8eebf4349011" [ 1016.069065] env[62952]: _type = "Task" [ 1016.069065] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.078224] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f54915-ccdc-37dd-5a26-8eebf4349011, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.087732] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.088313] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.090983] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.679s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.090983] env[62952]: DEBUG nova.objects.instance [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lazy-loading 'resources' on Instance uuid 4c4845e6-79ca-4a2a-a636-842d6919e70f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.145899] env[62952]: DEBUG oslo_vmware.api [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367468, 'name': PowerOnVM_Task, 'duration_secs': 0.813843} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.146260] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.146605] env[62952]: INFO nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Took 12.51 seconds to spawn the instance on the hypervisor. [ 1016.146933] env[62952]: DEBUG nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.147776] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d55489-5872-4820-9622-b3679ba19158 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.176443] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137b0368-2d8d-4b60-9c94-d8a8107518aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.199029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c820a38b-a398-42ec-b592-b7335160c239 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.362046] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fb63d7-ca9b-6a90-a860-8877f94a2caa, 'name': SearchDatastore_Task, 'duration_secs': 0.013149} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.362046] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.362046] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1016.362321] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24a8165c-115f-4421-bc8f-cbb2b6146bf8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.364316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.372474] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1016.372474] env[62952]: value = "task-1367470" [ 1016.372474] env[62952]: _type = "Task" [ 1016.372474] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.381753] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.411247] env[62952]: DEBUG oslo_concurrency.lockutils [req-44cb9e33-52ab-4c7a-a688-49791a2951bd req-cb604cc5-6fdf-42ca-b7a0-f76989b64fe6 service nova] Releasing lock "refresh_cache-1df00ddb-64fd-4ad3-be02-072a5aaad6f2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.567065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "interface-303c9fe7-0d89-4a1a-adcc-6f787486df61-e1654a86-c311-45dd-9cc6-143cf350df90" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.567480] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-303c9fe7-0d89-4a1a-adcc-6f787486df61-e1654a86-c311-45dd-9cc6-143cf350df90" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.567885] env[62952]: DEBUG nova.objects.instance [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lazy-loading 'flavor' on Instance uuid 303c9fe7-0d89-4a1a-adcc-6f787486df61 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.582613] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f54915-ccdc-37dd-5a26-8eebf4349011, 'name': SearchDatastore_Task, 'duration_secs': 0.012884} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.583771] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.584126] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.584368] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.584526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.584742] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.585547] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a84fd9f4-edf9-49bf-87bb-15e4814e1390 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.596019] env[62952]: DEBUG nova.compute.utils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.602181] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.602308] env[62952]: DEBUG nova.network.neutron [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1016.604765] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.604765] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.605439] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67de8409-55b1-4fcc-a2cb-18071b90afcb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.615224] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1016.615224] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52fe76cf-e561-36b4-445a-eea406b5d406" [ 1016.615224] env[62952]: _type = "Task" [ 1016.615224] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.625418] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fe76cf-e561-36b4-445a-eea406b5d406, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.652531] env[62952]: DEBUG nova.policy [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55ffb58a3b3a4ad29699d48ecd6c0700', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d564dc3c75e430dbaf1f90a9c90c18b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.669038] env[62952]: INFO nova.compute.manager [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Took 31.17 seconds to build instance. [ 1016.712393] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1016.712997] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a8be5cc8-1851-42ac-8864-1a977da5c501 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.725033] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1016.725033] env[62952]: value = "task-1367471" [ 1016.725033] env[62952]: _type = "Task" [ 1016.725033] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.739331] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367471, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.867087] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4378f0-0cd1-426e-862e-dbeb6cd85643 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.868970] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c80d0e03-68dc-4735-8d73-72f4bcf09450 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-ff189c9e-9e7c-4217-9c65-0f821393870a-e1654a86-c311-45dd-9cc6-143cf350df90" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.138s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.880752] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e054f57-248a-475c-94d1-1f8067cc1465 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.890786] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367470, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.922081] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba82be7-f635-4be7-9a84-06828ffe6bd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.931174] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedd9ae9-f87a-43f3-b5d8-6eab44bc0fc2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.952252] env[62952]: DEBUG nova.compute.provider_tree [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.979089] env[62952]: DEBUG nova.network.neutron [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Successfully created port: dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.101333] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.131451] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fe76cf-e561-36b4-445a-eea406b5d406, 'name': SearchDatastore_Task, 'duration_secs': 0.058969} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.132320] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e7442b7-235c-4709-8a39-e18f89360f08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.138748] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1017.138748] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52579895-24f8-1950-df28-dc7552edda2a" [ 1017.138748] env[62952]: _type = "Task" [ 1017.138748] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.149693] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52579895-24f8-1950-df28-dc7552edda2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.172365] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a408b6ce-9f18-4167-8678-e3fb72d21343 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.680s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.234650] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367471, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.242104] env[62952]: DEBUG nova.objects.instance [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lazy-loading 'pci_requests' on Instance uuid 303c9fe7-0d89-4a1a-adcc-6f787486df61 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.385981] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367470, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.393016] env[62952]: DEBUG nova.compute.manager [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-changed-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.393016] env[62952]: DEBUG nova.compute.manager [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing instance network info cache due to event network-changed-48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.393016] env[62952]: DEBUG oslo_concurrency.lockutils [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] Acquiring lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.393016] env[62952]: DEBUG oslo_concurrency.lockutils [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] Acquired lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.393315] env[62952]: DEBUG nova.network.neutron [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Refreshing network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.454331] env[62952]: DEBUG nova.scheduler.client.report [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.650125] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52579895-24f8-1950-df28-dc7552edda2a, 'name': SearchDatastore_Task, 'duration_secs': 0.043328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.650405] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.650664] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 1df00ddb-64fd-4ad3-be02-072a5aaad6f2/1df00ddb-64fd-4ad3-be02-072a5aaad6f2.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.650939] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6617d617-8eab-4c5b-8651-f7e02aca3acd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.658273] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1017.658273] env[62952]: value = "task-1367472" [ 1017.658273] env[62952]: _type = "Task" [ 1017.658273] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.666675] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367472, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.734116] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367471, 'name': CreateSnapshot_Task, 'duration_secs': 0.942455} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.734390] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1017.735193] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4adc479-886b-4557-88d1-f2abbbbba9b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.745072] env[62952]: DEBUG nova.objects.base [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Object Instance<303c9fe7-0d89-4a1a-adcc-6f787486df61> lazy-loaded attributes: flavor,pci_requests {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1017.745279] env[62952]: DEBUG nova.network.neutron [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1017.820540] env[62952]: DEBUG nova.policy [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb541b4abd9426ea96edcd066c69be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d1f373267464f14ae2c833151821973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1017.885950] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367470, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.959174] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.868s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.961980] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.357s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.962423] env[62952]: DEBUG nova.objects.instance [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lazy-loading 'resources' on Instance uuid ba2dfb85-0ffa-4883-abf1-482441c5afb8 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.992701] env[62952]: INFO nova.scheduler.client.report [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted allocations for instance 4c4845e6-79ca-4a2a-a636-842d6919e70f [ 1018.111287] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.144824] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5d1699f9bbd68d1b5a8ac81250869129',container_format='bare',created_at=2024-10-26T01:40:41Z,direct_url=,disk_format='vmdk',id=1de1bca9-aeab-4b87-85e8-152a9124e537,min_disk=1,min_ram=0,name='tempest-test-snap-971998641',owner='6d564dc3c75e430dbaf1f90a9c90c18b',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-26T01:40:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.145132] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.145270] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.145459] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.145605] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.145784] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.146026] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.146198] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.146372] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.146608] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.146823] env[62952]: DEBUG nova.virt.hardware [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.147780] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6034a9f-e4b1-4ea3-b89f-945cfb5b6558 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.157614] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ef6675-d875-47f4-883f-68811d78bcc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.171983] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367472, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.243143] env[62952]: DEBUG nova.network.neutron [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updated VIF entry in instance network info cache for port 48403971-25e5-4da0-a350-bd47ba090afc. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.243327] env[62952]: DEBUG nova.network.neutron [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [{"id": "48403971-25e5-4da0-a350-bd47ba090afc", "address": "fa:16:3e:b9:99:64", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48403971-25", "ovs_interfaceid": "48403971-25e5-4da0-a350-bd47ba090afc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.255141] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1018.256163] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2b3adc68-e13e-4ff5-8ed7-027aaf275f5b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.267516] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1018.267516] env[62952]: value = "task-1367473" [ 1018.267516] env[62952]: _type = "Task" [ 1018.267516] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.277551] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367473, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.389427] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367470, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.535434} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.389762] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1018.389982] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.390235] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0e11870-2b9f-4ae9-82ff-5e0ea1f168ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.398819] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1018.398819] env[62952]: value = "task-1367474" [ 1018.398819] env[62952]: _type = "Task" [ 1018.398819] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.409661] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.507154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e801fc0f-5bee-45f6-9b36-13536a0a6f9c tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "4c4845e6-79ca-4a2a-a636-842d6919e70f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.117s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.567769] env[62952]: DEBUG nova.compute.manager [req-d0e57816-f36c-4975-9a83-eee40d036320 req-e0433c4a-0151-4318-bdff-a3e2a1ccdac5 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Received event network-vif-plugged-dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.568023] env[62952]: DEBUG oslo_concurrency.lockutils [req-d0e57816-f36c-4975-9a83-eee40d036320 req-e0433c4a-0151-4318-bdff-a3e2a1ccdac5 service nova] Acquiring lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.568325] env[62952]: DEBUG oslo_concurrency.lockutils [req-d0e57816-f36c-4975-9a83-eee40d036320 req-e0433c4a-0151-4318-bdff-a3e2a1ccdac5 service nova] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.568536] env[62952]: DEBUG oslo_concurrency.lockutils [req-d0e57816-f36c-4975-9a83-eee40d036320 req-e0433c4a-0151-4318-bdff-a3e2a1ccdac5 service nova] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.568751] env[62952]: DEBUG nova.compute.manager [req-d0e57816-f36c-4975-9a83-eee40d036320 req-e0433c4a-0151-4318-bdff-a3e2a1ccdac5 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] No waiting events found dispatching network-vif-plugged-dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.568954] env[62952]: WARNING nova.compute.manager [req-d0e57816-f36c-4975-9a83-eee40d036320 req-e0433c4a-0151-4318-bdff-a3e2a1ccdac5 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Received unexpected event network-vif-plugged-dfb9de49-97fd-4f78-8770-4ac0e79bd8ed for instance with vm_state building and task_state spawning. [ 1018.646974] env[62952]: DEBUG nova.network.neutron [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Successfully updated port: dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.675879] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367472, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.729328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.678611] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 1df00ddb-64fd-4ad3-be02-072a5aaad6f2/1df00ddb-64fd-4ad3-be02-072a5aaad6f2.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1018.678916] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.679425] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-642a2fd2-2fae-4261-b851-724eaff0bcdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.687805] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1018.687805] env[62952]: value = "task-1367475" [ 1018.687805] env[62952]: _type = "Task" [ 1018.687805] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.693422] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0657668f-613b-4791-ab2b-e58c2961f380 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.703966] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367475, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.706997] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cbc150-648a-40ca-8a90-6911e7434da6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.741022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc72f1c2-6893-4525-aab4-5f01f861c5f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.745609] env[62952]: DEBUG oslo_concurrency.lockutils [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] Releasing lock "refresh_cache-ff189c9e-9e7c-4217-9c65-0f821393870a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.745897] env[62952]: DEBUG nova.compute.manager [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.746131] env[62952]: DEBUG nova.compute.manager [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing instance network info cache due to event network-changed-3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.746403] env[62952]: DEBUG oslo_concurrency.lockutils [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.746579] env[62952]: DEBUG oslo_concurrency.lockutils [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.746842] env[62952]: DEBUG nova.network.neutron [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.751689] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cd4e5a-4114-48ab-ae70-a0f5260a6a5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.773299] env[62952]: DEBUG nova.compute.provider_tree [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.786068] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367473, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.909229] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081741} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.909432] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.910294] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30b59cc-84d0-487f-a84b-fd9a57ca23c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.932463] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.932728] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5677f776-fe1a-4e19-9658-7bfe4fd59167 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.952969] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1018.952969] env[62952]: value = "task-1367476" [ 1018.952969] env[62952]: _type = "Task" [ 1018.952969] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.963802] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367476, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.149798] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "refresh_cache-e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.149798] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "refresh_cache-e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.149996] env[62952]: DEBUG nova.network.neutron [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.198134] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367475, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069185} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.198421] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1019.199323] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a8b8de-91ff-4493-b215-9345d5263922 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.222740] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 1df00ddb-64fd-4ad3-be02-072a5aaad6f2/1df00ddb-64fd-4ad3-be02-072a5aaad6f2.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.223088] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d234aab-9fe7-464a-aecd-7219b0a60067 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.245516] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1019.245516] env[62952]: value = "task-1367477" [ 1019.245516] env[62952]: _type = "Task" [ 1019.245516] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.256602] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367477, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.267746] env[62952]: DEBUG nova.compute.manager [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1019.283200] env[62952]: DEBUG nova.scheduler.client.report [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.292522] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367473, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.433236] env[62952]: DEBUG nova.network.neutron [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Successfully updated port: e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1019.467711] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367476, 'name': ReconfigVM_Task, 'duration_secs': 0.473762} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.468152] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f/1e02f16d-b86e-4533-928b-a672ce69788f.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.468899] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97a09e1a-8438-4aeb-acea-683446be19ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.477781] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1019.477781] env[62952]: value = "task-1367478" [ 1019.477781] env[62952]: _type = "Task" [ 1019.477781] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.487175] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367478, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.560301] env[62952]: DEBUG nova.network.neutron [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updated VIF entry in instance network info cache for port 3a39f7b2-d402-4ead-8172-34ae1a9c4387. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.560781] env[62952]: DEBUG nova.network.neutron [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.680851] env[62952]: DEBUG nova.network.neutron [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1019.756034] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367477, 'name': ReconfigVM_Task, 'duration_secs': 0.392433} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.756353] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 1df00ddb-64fd-4ad3-be02-072a5aaad6f2/1df00ddb-64fd-4ad3-be02-072a5aaad6f2.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.757019] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-595868bb-939a-4bcf-a33b-4039e8211617 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.765586] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1019.765586] env[62952]: value = "task-1367479" [ 1019.765586] env[62952]: _type = "Task" [ 1019.765586] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.778514] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367479, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.787556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.789533] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367473, 'name': CloneVM_Task, 'duration_secs': 1.369986} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.790185] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.338s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.792020] env[62952]: INFO nova.compute.claims [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.794901] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created linked-clone VM from snapshot [ 1019.796307] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.797754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485556c8-5489-454d-8b22-4686ca3b5bb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.808553] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Uploading image b30cf8b0-879c-40dd-8e67-0af16c12567b {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1019.811318] env[62952]: INFO nova.scheduler.client.report [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Deleted allocations for instance ba2dfb85-0ffa-4883-abf1-482441c5afb8 [ 1019.815444] env[62952]: DEBUG nova.network.neutron [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Updating instance_info_cache with network_info: [{"id": "dfb9de49-97fd-4f78-8770-4ac0e79bd8ed", "address": "fa:16:3e:e9:73:b6", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb9de49-97", "ovs_interfaceid": "dfb9de49-97fd-4f78-8770-4ac0e79bd8ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.833099] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1019.833099] env[62952]: value = "vm-290959" [ 1019.833099] env[62952]: _type = "VirtualMachine" [ 1019.833099] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1019.833625] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-510eef38-a59c-4fd6-90fe-3e47fe667420 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.842318] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease: (returnval){ [ 1019.842318] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b51036-5bcf-a9e2-4142-dc4c12478d81" [ 1019.842318] env[62952]: _type = "HttpNfcLease" [ 1019.842318] env[62952]: } obtained for exporting VM: (result){ [ 1019.842318] env[62952]: value = "vm-290959" [ 1019.842318] env[62952]: _type = "VirtualMachine" [ 1019.842318] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1019.842653] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the lease: (returnval){ [ 1019.842653] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b51036-5bcf-a9e2-4142-dc4c12478d81" [ 1019.842653] env[62952]: _type = "HttpNfcLease" [ 1019.842653] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1019.850677] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1019.850677] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b51036-5bcf-a9e2-4142-dc4c12478d81" [ 1019.850677] env[62952]: _type = "HttpNfcLease" [ 1019.850677] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1019.938622] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.989319] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367478, 'name': Rename_Task, 'duration_secs': 0.19424} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.989606] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.989872] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c69384e8-3d9d-46a6-a0d5-c7d85a628851 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.996155] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1019.996155] env[62952]: value = "task-1367481" [ 1019.996155] env[62952]: _type = "Task" [ 1019.996155] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.003570] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367481, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.063447] env[62952]: DEBUG oslo_concurrency.lockutils [req-7538a9f9-6abc-424d-a992-dbaf8adfe8af req-4e0a4935-4777-4ce3-a07b-7acfdb1977ab service nova] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.064078] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.064173] env[62952]: DEBUG nova.network.neutron [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.276380] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367479, 'name': Rename_Task, 'duration_secs': 0.157686} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.276558] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.276751] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8da18dec-a3c4-4ab1-a74d-fcd0837eeff6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.283754] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1020.283754] env[62952]: value = "task-1367482" [ 1020.283754] env[62952]: _type = "Task" [ 1020.283754] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.291748] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.320065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "refresh_cache-e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.320406] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Instance network_info: |[{"id": "dfb9de49-97fd-4f78-8770-4ac0e79bd8ed", "address": "fa:16:3e:e9:73:b6", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb9de49-97", "ovs_interfaceid": "dfb9de49-97fd-4f78-8770-4ac0e79bd8ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.321405] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:73:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92fe29b3-0907-453d-aabb-5559c4bd7c0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfb9de49-97fd-4f78-8770-4ac0e79bd8ed', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.329090] env[62952]: DEBUG oslo.service.loopingcall [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.329573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-12dccaae-9f5f-44a2-85d1-0e37c9ed33ce tempest-MultipleCreateTestJSON-878420833 tempest-MultipleCreateTestJSON-878420833-project-member] Lock "ba2dfb85-0ffa-4883-abf1-482441c5afb8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.632s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.331253] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.331504] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1c614c4-5223-47d8-9c82-f8e2be5b1236 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.354682] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1020.354682] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b51036-5bcf-a9e2-4142-dc4c12478d81" [ 1020.354682] env[62952]: _type = "HttpNfcLease" [ 1020.354682] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1020.355933] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1020.355933] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b51036-5bcf-a9e2-4142-dc4c12478d81" [ 1020.355933] env[62952]: _type = "HttpNfcLease" [ 1020.355933] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1020.356191] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.356191] env[62952]: value = "task-1367483" [ 1020.356191] env[62952]: _type = "Task" [ 1020.356191] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.356919] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca03dd5-6322-4c2e-87fd-b624eccb6819 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.367319] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367483, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.370702] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5257b372-b8c0-3201-30ed-336cdef5ac96/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1020.370917] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5257b372-b8c0-3201-30ed-336cdef5ac96/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1020.473163] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4b5cec22-5a49-4b5a-af1d-d1a39453fb8b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.511878] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367481, 'name': PowerOnVM_Task, 'duration_secs': 0.504901} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.512277] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.512548] env[62952]: DEBUG nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.513464] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0125684f-77ed-4db4-98d2-6a11054d105b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.601730] env[62952]: WARNING nova.network.neutron [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] cab59c10-e233-4404-bbf5-40841bb12dbd already exists in list: networks containing: ['cab59c10-e233-4404-bbf5-40841bb12dbd']. ignoring it [ 1020.645522] env[62952]: DEBUG nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Received event network-changed-dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.645858] env[62952]: DEBUG nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Refreshing instance network info cache due to event network-changed-dfb9de49-97fd-4f78-8770-4ac0e79bd8ed. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.646048] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Acquiring lock "refresh_cache-e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.646305] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Acquired lock "refresh_cache-e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.646676] env[62952]: DEBUG nova.network.neutron [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Refreshing network info cache for port dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.795990] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367482, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.872225] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367483, 'name': CreateVM_Task, 'duration_secs': 0.399094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.872486] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.873385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.873627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.874153] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.874515] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d79e76c-8c15-4670-96d5-27775d743072 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.884810] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1020.884810] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523194aa-c64e-d552-b6fa-76e0efe3f751" [ 1020.884810] env[62952]: _type = "Task" [ 1020.884810] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.894477] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523194aa-c64e-d552-b6fa-76e0efe3f751, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.024262] env[62952]: INFO nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] bringing vm to original state: 'stopped' [ 1021.075705] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b98f4f-7aea-407b-894d-a00cf95b424f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.088522] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5edfada-745f-4162-9339-dd8532b2f454 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.126404] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3735fbb-25d1-4fc7-b6d1-f39034acab9c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.134689] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f3ad38-4ac9-4fb6-a458-1ccb2137967a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.153076] env[62952]: DEBUG nova.compute.provider_tree [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.181200] env[62952]: DEBUG nova.network.neutron [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1654a86-c311-45dd-9cc6-143cf350df90", "address": "fa:16:3e:2d:97:e7", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1654a86-c3", "ovs_interfaceid": "e1654a86-c311-45dd-9cc6-143cf350df90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.296698] env[62952]: DEBUG oslo_vmware.api [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367482, 'name': PowerOnVM_Task, 'duration_secs': 0.538457} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.299845] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.300128] env[62952]: INFO nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1021.300606] env[62952]: DEBUG nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.301506] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615efc10-829c-4909-b2cb-155d500414a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.381470] env[62952]: DEBUG nova.network.neutron [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Updated VIF entry in instance network info cache for port dfb9de49-97fd-4f78-8770-4ac0e79bd8ed. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.381827] env[62952]: DEBUG nova.network.neutron [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Updating instance_info_cache with network_info: [{"id": "dfb9de49-97fd-4f78-8770-4ac0e79bd8ed", "address": "fa:16:3e:e9:73:b6", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb9de49-97", "ovs_interfaceid": "dfb9de49-97fd-4f78-8770-4ac0e79bd8ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.399331] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.399696] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Processing image 1de1bca9-aeab-4b87-85e8-152a9124e537 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.399997] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.400229] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.400377] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.400727] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14f82c23-c0cf-4487-9e0f-29bbdb980614 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.420253] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.420563] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.421782] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-717999d6-9a0b-4ccd-8f32-153746a3b0fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.428030] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1021.428030] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52dd3c73-3712-69ff-15ed-06a14369c391" [ 1021.428030] env[62952]: _type = "Task" [ 1021.428030] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.437900] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52dd3c73-3712-69ff-15ed-06a14369c391, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.656633] env[62952]: DEBUG nova.scheduler.client.report [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.683467] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.684762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.685192] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.686149] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3661c40-7c5b-4620-8cca-777329b45f2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.705420] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.705880] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.706483] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.706841] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.707167] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.707583] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.707999] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.708388] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.712017] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.712017] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.712017] env[62952]: DEBUG nova.virt.hardware [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.716390] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Reconfiguring VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1021.716897] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d75219a1-8bad-4c2e-8750-815859137c3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.738639] env[62952]: DEBUG oslo_vmware.api [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1021.738639] env[62952]: value = "task-1367484" [ 1021.738639] env[62952]: _type = "Task" [ 1021.738639] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.748382] env[62952]: DEBUG oslo_vmware.api [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367484, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.823141] env[62952]: INFO nova.compute.manager [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Took 20.73 seconds to build instance. [ 1021.886026] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Releasing lock "refresh_cache-e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.886026] env[62952]: DEBUG nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-vif-plugged-e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.886168] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.886282] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.886448] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.886642] env[62952]: DEBUG nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] No waiting events found dispatching network-vif-plugged-e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.886909] env[62952]: WARNING nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received unexpected event network-vif-plugged-e1654a86-c311-45dd-9cc6-143cf350df90 for instance with vm_state active and task_state None. [ 1021.887220] env[62952]: DEBUG nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-changed-e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.887416] env[62952]: DEBUG nova.compute.manager [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing instance network info cache due to event network-changed-e1654a86-c311-45dd-9cc6-143cf350df90. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1021.887625] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.887819] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.887948] env[62952]: DEBUG nova.network.neutron [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Refreshing network info cache for port e1654a86-c311-45dd-9cc6-143cf350df90 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.939896] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1021.940295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Fetch image to [datastore1] OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247/OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1021.940524] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Downloading stream optimized image 1de1bca9-aeab-4b87-85e8-152a9124e537 to [datastore1] OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247/OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247.vmdk on the data store datastore1 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1021.940779] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Downloading image file data 1de1bca9-aeab-4b87-85e8-152a9124e537 to the ESX as VM named 'OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1022.031340] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1022.031340] env[62952]: value = "resgroup-9" [ 1022.031340] env[62952]: _type = "ResourcePool" [ 1022.031340] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1022.035015] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c52afeb8-28ed-464b-9e80-1aa938400357 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.051232] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "1e02f16d-b86e-4533-928b-a672ce69788f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.051618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.051900] env[62952]: DEBUG nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.052866] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dba947e-fa4e-4ca8-99c7-ffe40bbf6ffe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.061033] env[62952]: DEBUG nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1022.063970] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.063970] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f64ca7aa-d44c-4ffb-b4c7-15935d8b30f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.066994] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lease: (returnval){ [ 1022.066994] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52edf8d8-f228-cee4-7ae5-448366ae0d60" [ 1022.066994] env[62952]: _type = "HttpNfcLease" [ 1022.066994] env[62952]: } obtained for vApp import into resource pool (val){ [ 1022.066994] env[62952]: value = "resgroup-9" [ 1022.066994] env[62952]: _type = "ResourcePool" [ 1022.066994] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1022.067981] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the lease: (returnval){ [ 1022.067981] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52edf8d8-f228-cee4-7ae5-448366ae0d60" [ 1022.067981] env[62952]: _type = "HttpNfcLease" [ 1022.067981] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1022.072807] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1022.072807] env[62952]: value = "task-1367486" [ 1022.072807] env[62952]: _type = "Task" [ 1022.072807] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.074587] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1022.074587] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52edf8d8-f228-cee4-7ae5-448366ae0d60" [ 1022.074587] env[62952]: _type = "HttpNfcLease" [ 1022.074587] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1022.086491] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367486, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.163745] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.164579] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1022.168106] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.168567] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.547s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.168859] env[62952]: DEBUG nova.objects.instance [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lazy-loading 'resources' on Instance uuid 82401449-347e-4917-a307-d5d7bc048f7c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.252274] env[62952]: DEBUG oslo_vmware.api [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.326035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a01be3a9-0f25-4e00-95fc-85e151527d8e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.240s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.327592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.158s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.327592] env[62952]: DEBUG nova.compute.manager [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.328320] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4949cd-0b90-42ac-9608-82667a0303a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.334676] env[62952]: DEBUG nova.compute.manager [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1022.335436] env[62952]: DEBUG nova.objects.instance [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lazy-loading 'flavor' on Instance uuid 1df00ddb-64fd-4ad3-be02-072a5aaad6f2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.577843] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1022.577843] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52edf8d8-f228-cee4-7ae5-448366ae0d60" [ 1022.577843] env[62952]: _type = "HttpNfcLease" [ 1022.577843] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1022.587080] env[62952]: DEBUG oslo_vmware.api [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367486, 'name': PowerOffVM_Task, 'duration_secs': 0.309077} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.589724] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.589952] env[62952]: DEBUG nova.compute.manager [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.590793] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d481398-de4b-45a3-9ef0-7a4b96df38b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.633838] env[62952]: DEBUG nova.network.neutron [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updated VIF entry in instance network info cache for port e1654a86-c311-45dd-9cc6-143cf350df90. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1022.633838] env[62952]: DEBUG nova.network.neutron [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1654a86-c311-45dd-9cc6-143cf350df90", "address": "fa:16:3e:2d:97:e7", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1654a86-c3", "ovs_interfaceid": "e1654a86-c311-45dd-9cc6-143cf350df90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.672929] env[62952]: DEBUG nova.compute.utils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.677422] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1022.677933] env[62952]: DEBUG nova.network.neutron [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1022.719666] env[62952]: DEBUG nova.policy [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.757752] env[62952]: DEBUG oslo_vmware.api [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367484, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.843459] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.843912] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab3cce7f-7b38-4e35-8ce4-96fbbf92a7eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.851190] env[62952]: DEBUG oslo_vmware.api [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1022.851190] env[62952]: value = "task-1367487" [ 1022.851190] env[62952]: _type = "Task" [ 1022.851190] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.860784] env[62952]: DEBUG oslo_vmware.api [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.917942] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7848cb3a-ab5b-4af1-a570-ef9a2cff15e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.926167] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c178eee-3ee0-42ba-bac8-37da235d54a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.958268] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7a1827-e84b-4bc9-ab0a-3d0934e1f941 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.966593] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2cefc9-f633-4f32-98c3-e48bb7f32967 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.980900] env[62952]: DEBUG nova.compute.provider_tree [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.036627] env[62952]: DEBUG nova.network.neutron [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Successfully created port: 8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.078524] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1023.078524] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52edf8d8-f228-cee4-7ae5-448366ae0d60" [ 1023.078524] env[62952]: _type = "HttpNfcLease" [ 1023.078524] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1023.078958] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1023.078958] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52edf8d8-f228-cee4-7ae5-448366ae0d60" [ 1023.078958] env[62952]: _type = "HttpNfcLease" [ 1023.078958] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1023.079528] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bfd408-72dc-4df6-96a6-2a5dcdd036c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.088947] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b3f55-abe0-60a1-c08d-e1a3684fe31a/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1023.089240] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b3f55-abe0-60a1-c08d-e1a3684fe31a/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1023.158353] env[62952]: DEBUG oslo_concurrency.lockutils [req-8e87ff8d-520e-44e9-ba1e-0ea5fb0f5fca req-1aa49529-4d5b-4291-896a-0352216e4e73 service nova] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.159013] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.107s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.167266] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-317a9460-2a9e-4d85-bdba-b9263a830688 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.175245] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.252360] env[62952]: DEBUG oslo_vmware.api [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367484, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.364510] env[62952]: DEBUG oslo_vmware.api [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367487, 'name': PowerOffVM_Task, 'duration_secs': 0.201893} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.364810] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.364993] env[62952]: DEBUG nova.compute.manager [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.365824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b6183e-878e-4961-ac30-722a403d67ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.484301] env[62952]: DEBUG nova.scheduler.client.report [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.672401] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.754343] env[62952]: DEBUG oslo_vmware.api [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367484, 'name': ReconfigVM_Task, 'duration_secs': 1.94411} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.756412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.756766] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Reconfigured VM to attach interface {{(pid=62952) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1023.879747] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e44d68a5-6929-41b2-8c9a-88a343b20784 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.553s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.960633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "1e02f16d-b86e-4533-928b-a672ce69788f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.960633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.960633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "1e02f16d-b86e-4533-928b-a672ce69788f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.960633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.960633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.963481] env[62952]: INFO nova.compute.manager [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Terminating instance [ 1023.965432] env[62952]: DEBUG nova.compute.manager [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.965637] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.966515] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846cbbea-0f44-44dd-9ead-12842a1d609c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.983604] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.984049] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb4c8c8b-9bbd-45f1-8dce-1258361203f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.988997] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.992828] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.197s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.016465] env[62952]: INFO nova.scheduler.client.report [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Deleted allocations for instance 82401449-347e-4917-a307-d5d7bc048f7c [ 1024.081113] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1024.081516] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1024.081669] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore1] 1e02f16d-b86e-4533-928b-a672ce69788f {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.081982] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0aae3978-5575-4759-bc81-fc60dcea6519 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.092090] env[62952]: DEBUG oslo_vmware.api [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1024.092090] env[62952]: value = "task-1367489" [ 1024.092090] env[62952]: _type = "Task" [ 1024.092090] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.103029] env[62952]: DEBUG oslo_vmware.api [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.189581] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1024.215083] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.215436] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.215641] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.215827] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.215987] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.216156] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.216403] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.216610] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.216832] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.217098] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.217364] env[62952]: DEBUG nova.virt.hardware [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.218456] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bbcee5-bc11-415d-9c37-edc3a1c70a0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.232744] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c284c7-79c2-4e6c-a480-ca1ef0dccf46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.262208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74dbd872-5536-437e-8ed0-e663086f2e0b tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-303c9fe7-0d89-4a1a-adcc-6f787486df61-e1654a86-c311-45dd-9cc6-143cf350df90" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.695s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.289225] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1024.289448] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b3f55-abe0-60a1-c08d-e1a3684fe31a/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1024.290480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60651173-e0b0-4660-aff3-91401e511ebd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.298081] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b3f55-abe0-60a1-c08d-e1a3684fe31a/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1024.298305] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b3f55-abe0-60a1-c08d-e1a3684fe31a/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1024.298560] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7656e326-4cb3-45f0-bb21-27721e9169b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.500791] env[62952]: INFO nova.compute.claims [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.525780] env[62952]: DEBUG oslo_concurrency.lockutils [None req-db6c8fa2-3099-4273-90d8-75d39878836d tempest-ServerShowV247Test-1639492440 tempest-ServerShowV247Test-1639492440-project-member] Lock "82401449-347e-4917-a307-d5d7bc048f7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.656s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.561039] env[62952]: DEBUG nova.network.neutron [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Successfully updated port: 8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1024.590028] env[62952]: DEBUG nova.compute.manager [req-80c5f525-a5f7-4a27-b763-a7c173bd317a req-717bd320-dac1-4e85-af40-6b7b185226db service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Received event network-vif-plugged-8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.590362] env[62952]: DEBUG oslo_concurrency.lockutils [req-80c5f525-a5f7-4a27-b763-a7c173bd317a req-717bd320-dac1-4e85-af40-6b7b185226db service nova] Acquiring lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.590656] env[62952]: DEBUG oslo_concurrency.lockutils [req-80c5f525-a5f7-4a27-b763-a7c173bd317a req-717bd320-dac1-4e85-af40-6b7b185226db service nova] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.590939] env[62952]: DEBUG oslo_concurrency.lockutils [req-80c5f525-a5f7-4a27-b763-a7c173bd317a req-717bd320-dac1-4e85-af40-6b7b185226db service nova] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.591156] env[62952]: DEBUG nova.compute.manager [req-80c5f525-a5f7-4a27-b763-a7c173bd317a req-717bd320-dac1-4e85-af40-6b7b185226db service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] No waiting events found dispatching network-vif-plugged-8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1024.591340] env[62952]: WARNING nova.compute.manager [req-80c5f525-a5f7-4a27-b763-a7c173bd317a req-717bd320-dac1-4e85-af40-6b7b185226db service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Received unexpected event network-vif-plugged-8514c98b-afbf-4e2a-bccf-d33990628a2e for instance with vm_state building and task_state spawning. [ 1024.605970] env[62952]: DEBUG oslo_vmware.api [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176201} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.606403] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.606674] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.607052] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.607392] env[62952]: INFO nova.compute.manager [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1024.607822] env[62952]: DEBUG oslo.service.loopingcall [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.608608] env[62952]: DEBUG nova.compute.manager [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1024.608754] env[62952]: DEBUG nova.network.neutron [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1024.704832] env[62952]: DEBUG oslo_vmware.rw_handles [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b3f55-abe0-60a1-c08d-e1a3684fe31a/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1024.705139] env[62952]: INFO nova.virt.vmwareapi.images [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Downloaded image file data 1de1bca9-aeab-4b87-85e8-152a9124e537 [ 1024.705979] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b038d7f-418d-4bda-aa8b-3813044e7031 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.723642] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77d2b7c4-24b2-47a5-aa92-e8c2064c71a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.752727] env[62952]: INFO nova.virt.vmwareapi.images [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] The imported VM was unregistered [ 1024.754243] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1024.754481] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating directory with path [datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.754793] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b466d144-18ab-47f5-b518-cbd9e531a4c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.770689] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created directory with path [datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.770689] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247/OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247.vmdk to [datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1024.770689] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d267b8e7-290c-4b41-a54e-ff1359a2f748 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.776524] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1024.776524] env[62952]: value = "task-1367491" [ 1024.776524] env[62952]: _type = "Task" [ 1024.776524] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.784325] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.963299] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.963764] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.964152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.964506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.964821] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.968425] env[62952]: INFO nova.compute.manager [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Terminating instance [ 1024.971574] env[62952]: DEBUG nova.compute.manager [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.971923] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.973276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1910de45-21c0-4760-82b0-1b916437a3b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.984883] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1024.985367] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6525f63-9dc4-4212-a82c-7fa715e24fed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.009355] env[62952]: INFO nova.compute.resource_tracker [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating resource usage from migration d1159953-ebe6-4466-a646-a5de8ba8569f [ 1025.063096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-2986a508-3f0b-4bad-b785-9e9f3ca7e39c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.063266] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-2986a508-3f0b-4bad-b785-9e9f3ca7e39c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.063427] env[62952]: DEBUG nova.network.neutron [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.105383] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.105671] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.105954] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleting the datastore file [datastore1] 1df00ddb-64fd-4ad3-be02-072a5aaad6f2 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.106323] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f10f03b-c0d5-40fb-8110-905902da4198 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.115028] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1025.115028] env[62952]: value = "task-1367493" [ 1025.115028] env[62952]: _type = "Task" [ 1025.115028] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.130616] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.228728] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2eb86b4-57de-4186-9193-e7d66fad4d24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.237481] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a315071-42f5-4694-898c-28289385d663 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.270365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b63394-31ae-4f69-a618-c442f6493fee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.283523] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d04691-a109-46ce-8b1b-85b3780e3112 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.298914] env[62952]: DEBUG nova.compute.provider_tree [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.303697] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.420353] env[62952]: DEBUG nova.network.neutron [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.598537] env[62952]: DEBUG nova.network.neutron [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1025.630186] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.789253] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.805588] env[62952]: DEBUG nova.scheduler.client.report [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.815169] env[62952]: DEBUG nova.network.neutron [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Updating instance_info_cache with network_info: [{"id": "8514c98b-afbf-4e2a-bccf-d33990628a2e", "address": "fa:16:3e:0b:ac:97", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8514c98b-af", "ovs_interfaceid": "8514c98b-afbf-4e2a-bccf-d33990628a2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.864715] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "interface-303c9fe7-0d89-4a1a-adcc-6f787486df61-e1654a86-c311-45dd-9cc6-143cf350df90" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.864996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-303c9fe7-0d89-4a1a-adcc-6f787486df61-e1654a86-c311-45dd-9cc6-143cf350df90" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.923661] env[62952]: INFO nova.compute.manager [-] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Took 1.31 seconds to deallocate network for instance. [ 1026.128947] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.289246] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.313029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.320s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.313029] env[62952]: INFO nova.compute.manager [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Migrating [ 1026.313029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.313029] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.314467] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.643s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.314890] env[62952]: DEBUG nova.objects.instance [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1026.317786] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-2986a508-3f0b-4bad-b785-9e9f3ca7e39c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.317786] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Instance network_info: |[{"id": "8514c98b-afbf-4e2a-bccf-d33990628a2e", "address": "fa:16:3e:0b:ac:97", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8514c98b-af", "ovs_interfaceid": "8514c98b-afbf-4e2a-bccf-d33990628a2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1026.318185] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:ac:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8514c98b-afbf-4e2a-bccf-d33990628a2e', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.326040] env[62952]: DEBUG oslo.service.loopingcall [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.326944] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1026.327218] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a6dc77d-8c2b-43b6-8db7-5223bd749cc5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.347968] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.347968] env[62952]: value = "task-1367494" [ 1026.347968] env[62952]: _type = "Task" [ 1026.347968] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.357296] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367494, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.367563] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.367773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.369235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258d22cc-b21c-42a8-a6ed-e5c41fb92c4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.387131] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ae5cf8-fb30-4043-aa89-f893c0f42197 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.414159] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Reconfiguring VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1026.414526] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-973948fe-c510-4e3e-a7cf-655e25f17f75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.430988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.434826] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1026.434826] env[62952]: value = "task-1367495" [ 1026.434826] env[62952]: _type = "Task" [ 1026.434826] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.444238] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.626081] env[62952]: DEBUG nova.compute.manager [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Received event network-changed-8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.626302] env[62952]: DEBUG nova.compute.manager [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Refreshing instance network info cache due to event network-changed-8514c98b-afbf-4e2a-bccf-d33990628a2e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1026.626820] env[62952]: DEBUG oslo_concurrency.lockutils [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] Acquiring lock "refresh_cache-2986a508-3f0b-4bad-b785-9e9f3ca7e39c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.626981] env[62952]: DEBUG oslo_concurrency.lockutils [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] Acquired lock "refresh_cache-2986a508-3f0b-4bad-b785-9e9f3ca7e39c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.627214] env[62952]: DEBUG nova.network.neutron [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Refreshing network info cache for port 8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.634815] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.788753] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.818447] env[62952]: INFO nova.compute.rpcapi [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 1026.819140] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.859063] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367494, 'name': CreateVM_Task, 'duration_secs': 0.413694} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.859353] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.860449] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.860639] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.861202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1026.861703] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93ce9de2-87af-4405-b4ab-0f7fe94be46b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.866619] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1026.866619] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526f922e-5207-3104-b459-667205b0c756" [ 1026.866619] env[62952]: _type = "Task" [ 1026.866619] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.874765] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526f922e-5207-3104-b459-667205b0c756, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.945727] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.128917] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.289757] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.337372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c0de5fa9-f106-451f-9fd8-5ad943348840 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.340163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.340163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.340163] env[62952]: DEBUG nova.network.neutron [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.341283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.910s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.341569] env[62952]: DEBUG nova.objects.instance [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'resources' on Instance uuid 1e02f16d-b86e-4533-928b-a672ce69788f {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.343564] env[62952]: DEBUG nova.network.neutron [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Updated VIF entry in instance network info cache for port 8514c98b-afbf-4e2a-bccf-d33990628a2e. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.344272] env[62952]: DEBUG nova.network.neutron [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Updating instance_info_cache with network_info: [{"id": "8514c98b-afbf-4e2a-bccf-d33990628a2e", "address": "fa:16:3e:0b:ac:97", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8514c98b-af", "ovs_interfaceid": "8514c98b-afbf-4e2a-bccf-d33990628a2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.379993] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526f922e-5207-3104-b459-667205b0c756, 'name': SearchDatastore_Task, 'duration_secs': 0.012465} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.380493] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.380833] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.381192] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.381427] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.381684] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.382074] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4934eeb-d23a-4aff-9688-d2f8b56f18bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.391272] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.391450] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1027.392189] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5919ff34-87d4-49f0-bcbc-e0cab125f52d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.399472] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1027.399472] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ade3f8-6db6-fc38-01f6-ebbe7f5bb667" [ 1027.399472] env[62952]: _type = "Task" [ 1027.399472] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.406909] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ade3f8-6db6-fc38-01f6-ebbe7f5bb667, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.445784] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.629599] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.791030] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.849319] env[62952]: DEBUG oslo_concurrency.lockutils [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] Releasing lock "refresh_cache-2986a508-3f0b-4bad-b785-9e9f3ca7e39c" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.849489] env[62952]: DEBUG nova.compute.manager [req-cf21a8a2-1749-4459-9187-8b630b992e07 req-2b3850b0-e759-46f9-bacb-76db2d6990a0 service nova] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Received event network-vif-deleted-97ceb54a-06d4-4154-9965-95093341cdb4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.912929] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ade3f8-6db6-fc38-01f6-ebbe7f5bb667, 'name': SearchDatastore_Task, 'duration_secs': 0.00947} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.913835] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-448ffc8d-3fc7-43b9-9e6b-14610c4af965 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.922584] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1027.922584] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ee6c35-84fa-22e1-cc1d-829f706866fc" [ 1027.922584] env[62952]: _type = "Task" [ 1027.922584] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.931667] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ee6c35-84fa-22e1-cc1d-829f706866fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.947221] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.094854] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bc2330-b8ab-45b5-9b68-1b0a450d5e0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.103222] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2995f60-e081-4131-a957-62aaac826abe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.141365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58e16e0-2727-47a2-aec0-5f272fb8e282 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.150529] env[62952]: DEBUG oslo_vmware.api [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.889076} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.152343] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.152615] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1028.152828] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1028.153046] env[62952]: INFO nova.compute.manager [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Took 3.18 seconds to destroy the instance on the hypervisor. [ 1028.153323] env[62952]: DEBUG oslo.service.loopingcall [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.153596] env[62952]: DEBUG nova.compute.manager [-] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.153691] env[62952]: DEBUG nova.network.neutron [-] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1028.156545] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd66530-5b08-4ce2-be0c-93a3bef92e5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.172965] env[62952]: DEBUG nova.compute.provider_tree [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.182190] env[62952]: DEBUG nova.network.neutron [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance_info_cache with network_info: [{"id": "3f88a864-36dc-4fc8-8073-977444f51ed8", "address": "fa:16:3e:70:82:3d", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f88a864-36", "ovs_interfaceid": "3f88a864-36dc-4fc8-8073-977444f51ed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.271942] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5257b372-b8c0-3201-30ed-336cdef5ac96/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1028.272913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6ff53c-9224-4873-b6f7-146b77decef7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.279481] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5257b372-b8c0-3201-30ed-336cdef5ac96/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1028.279660] env[62952]: ERROR oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5257b372-b8c0-3201-30ed-336cdef5ac96/disk-0.vmdk due to incomplete transfer. [ 1028.279882] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-52f6c989-c817-4828-8966-2c4ead258e9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.289128] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367491, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.483978} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.290137] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247/OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247.vmdk to [datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk. [ 1028.290415] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Cleaning up location [datastore1] OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1028.290588] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_02cbbd2e-83d6-41cc-b594-461bb825d247 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.290836] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5257b372-b8c0-3201-30ed-336cdef5ac96/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1028.291016] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Uploaded image b30cf8b0-879c-40dd-8e67-0af16c12567b to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1028.293350] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1028.293581] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-049cdfcc-2c3a-49ff-bc79-7bb50a8cba60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.295089] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-46943491-e6d4-44cb-bcd3-24ddfe822394 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.300378] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1028.300378] env[62952]: value = "task-1367497" [ 1028.300378] env[62952]: _type = "Task" [ 1028.300378] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.301488] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1028.301488] env[62952]: value = "task-1367496" [ 1028.301488] env[62952]: _type = "Task" [ 1028.301488] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.311338] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367496, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.314788] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367497, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.433931] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ee6c35-84fa-22e1-cc1d-829f706866fc, 'name': SearchDatastore_Task, 'duration_secs': 0.019529} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.434258] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.434543] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 2986a508-3f0b-4bad-b785-9e9f3ca7e39c/2986a508-3f0b-4bad-b785-9e9f3ca7e39c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.434821] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fbcfa3c-fff4-4927-b564-9ac956c220c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.442571] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1028.442571] env[62952]: value = "task-1367498" [ 1028.442571] env[62952]: _type = "Task" [ 1028.442571] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.449396] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.454148] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.648420] env[62952]: DEBUG nova.compute.manager [req-b791cb4e-9540-46a5-b4cb-1b10852f64f5 req-143ee376-73c0-435f-a9f0-6bfc35bc4ee8 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Received event network-vif-deleted-143465ea-10eb-4235-a204-f1a1003a57a6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.648628] env[62952]: INFO nova.compute.manager [req-b791cb4e-9540-46a5-b4cb-1b10852f64f5 req-143ee376-73c0-435f-a9f0-6bfc35bc4ee8 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Neutron deleted interface 143465ea-10eb-4235-a204-f1a1003a57a6; detaching it from the instance and deleting it from the info cache [ 1028.648807] env[62952]: DEBUG nova.network.neutron [req-b791cb4e-9540-46a5-b4cb-1b10852f64f5 req-143ee376-73c0-435f-a9f0-6bfc35bc4ee8 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.680336] env[62952]: DEBUG nova.scheduler.client.report [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.684267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.814101] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367496, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037983} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.817289] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.817481] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.817747] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk to [datastore1] e4fd9b9a-fc39-443c-9317-8dd286fdd2b0/e4fd9b9a-fc39-443c-9317-8dd286fdd2b0.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.818072] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367497, 'name': Destroy_Task, 'duration_secs': 0.35595} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.818587] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0de7cff-2a0e-41f7-b00e-6fa3e2f8db7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.820644] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroyed the VM [ 1028.821303] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1028.821303] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c7516db2-05ef-4110-8670-7b35e4a04b70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.828833] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1028.828833] env[62952]: value = "task-1367499" [ 1028.828833] env[62952]: _type = "Task" [ 1028.828833] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.830256] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1028.830256] env[62952]: value = "task-1367500" [ 1028.830256] env[62952]: _type = "Task" [ 1028.830256] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.843797] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367499, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.914159] env[62952]: DEBUG nova.network.neutron [-] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.950507] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.955735] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367498, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.152255] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e94aa8c-efaa-4a06-ae64-0198266614f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.161793] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bade65-2c69-4953-aaa9-92b9f3fea939 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.192108] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.198245] env[62952]: DEBUG nova.compute.manager [req-b791cb4e-9540-46a5-b4cb-1b10852f64f5 req-143ee376-73c0-435f-a9f0-6bfc35bc4ee8 service nova] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Detach interface failed, port_id=143465ea-10eb-4235-a204-f1a1003a57a6, reason: Instance 1df00ddb-64fd-4ad3-be02-072a5aaad6f2 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1029.228578] env[62952]: INFO nova.scheduler.client.report [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted allocations for instance 1e02f16d-b86e-4533-928b-a672ce69788f [ 1029.343311] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367499, 'name': RemoveSnapshot_Task} progress is 30%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.347177] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367500, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.417746] env[62952]: INFO nova.compute.manager [-] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Took 1.26 seconds to deallocate network for instance. [ 1029.451415] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.456817] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575065} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.457793] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 2986a508-3f0b-4bad-b785-9e9f3ca7e39c/2986a508-3f0b-4bad-b785-9e9f3ca7e39c.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.458032] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.458312] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9e3e69b-5205-407c-b289-2ceb696c344f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.464810] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1029.464810] env[62952]: value = "task-1367501" [ 1029.464810] env[62952]: _type = "Task" [ 1029.464810] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.473107] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.736801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a6d7a287-d5ce-49e8-b17b-bd19f4494ae5 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "1e02f16d-b86e-4533-928b-a672ce69788f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.777s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.844861] env[62952]: DEBUG oslo_vmware.api [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367499, 'name': RemoveSnapshot_Task, 'duration_secs': 0.613875} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.848018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1029.848290] env[62952]: INFO nova.compute.manager [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 13.67 seconds to snapshot the instance on the hypervisor. [ 1029.850767] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367500, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.925511] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.925792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.926035] env[62952]: DEBUG nova.objects.instance [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lazy-loading 'resources' on Instance uuid 1df00ddb-64fd-4ad3-be02-072a5aaad6f2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.949826] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.975733] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.318948} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.976057] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.977084] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9391f28-ce5d-48a6-92f7-2258ae918a02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.999521] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 2986a508-3f0b-4bad-b785-9e9f3ca7e39c/2986a508-3f0b-4bad-b785-9e9f3ca7e39c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.999793] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eba247f6-c2b7-4758-b9c6-9e1012e84d2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.022770] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1030.022770] env[62952]: value = "task-1367502" [ 1030.022770] env[62952]: _type = "Task" [ 1030.022770] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.031549] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367502, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.208489] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcee8ec-ac0d-48a7-8e80-f17c354f5441 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.227291] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.346975] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367500, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.398348] env[62952]: DEBUG nova.compute.manager [None req-be4bea76-cb05-4fbd-bbaa-aa66fd4edf65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Found 2 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1030.451091] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.533348] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367502, 'name': ReconfigVM_Task, 'duration_secs': 0.43209} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.534056] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 2986a508-3f0b-4bad-b785-9e9f3ca7e39c/2986a508-3f0b-4bad-b785-9e9f3ca7e39c.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.534795] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e71b1591-c293-499e-8b1f-ad46f2482bba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.544620] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1030.544620] env[62952]: value = "task-1367503" [ 1030.544620] env[62952]: _type = "Task" [ 1030.544620] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.557992] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367503, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.625033] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3606b277-4893-4df0-b7be-80d3cf846983 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.632824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653a323b-38bb-4bfe-92cd-745c7dd34460 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.665495] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3872700f-0749-4f95-a887-73b47668e6d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.673634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918f52d8-d00d-48bf-8266-ea6ca88339de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.688753] env[62952]: DEBUG nova.compute.provider_tree [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.733383] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.733718] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf099c4b-104e-4a64-a528-35bb586e8455 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.740941] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1030.740941] env[62952]: value = "task-1367504" [ 1030.740941] env[62952]: _type = "Task" [ 1030.740941] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.750667] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.847124] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367500, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.950972] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.055084] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367503, 'name': Rename_Task, 'duration_secs': 0.19417} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.055418] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.055687] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc422597-bea1-4f46-8c7d-c16e1d8b459b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.062796] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1031.062796] env[62952]: value = "task-1367505" [ 1031.062796] env[62952]: _type = "Task" [ 1031.062796] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.072974] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.192750] env[62952]: DEBUG nova.scheduler.client.report [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.210826] env[62952]: DEBUG nova.compute.manager [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.211972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6853f4b7-98b2-47d1-921f-911a7fb72a22 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.251622] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367504, 'name': PowerOffVM_Task, 'duration_secs': 0.423096} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.251913] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.252141] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1031.346614] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367500, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.390494} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.346893] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1de1bca9-aeab-4b87-85e8-152a9124e537/1de1bca9-aeab-4b87-85e8-152a9124e537.vmdk to [datastore1] e4fd9b9a-fc39-443c-9317-8dd286fdd2b0/e4fd9b9a-fc39-443c-9317-8dd286fdd2b0.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.347765] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e98a3ae-e3d8-480a-a102-2417f275eafc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.372014] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] e4fd9b9a-fc39-443c-9317-8dd286fdd2b0/e4fd9b9a-fc39-443c-9317-8dd286fdd2b0.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.372316] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6919ebb-6cd7-4e9b-bdf5-608e17eb4336 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.392943] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1031.392943] env[62952]: value = "task-1367506" [ 1031.392943] env[62952]: _type = "Task" [ 1031.392943] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.400722] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367506, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.450203] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.573229] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367505, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.701017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.725095] env[62952]: INFO nova.scheduler.client.report [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocations for instance 1df00ddb-64fd-4ad3-be02-072a5aaad6f2 [ 1031.725095] env[62952]: INFO nova.compute.manager [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] instance snapshotting [ 1031.725095] env[62952]: DEBUG nova.objects.instance [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'flavor' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.759203] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.759203] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.759487] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.759708] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.760209] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.760552] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.760728] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.761086] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.761572] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.761861] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.762397] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.768531] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9fd10ab-3db6-4735-ab20-0dd907e9f64a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.786411] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1031.786411] env[62952]: value = "task-1367507" [ 1031.786411] env[62952]: _type = "Task" [ 1031.786411] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.796056] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367507, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.904029] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367506, 'name': ReconfigVM_Task, 'duration_secs': 0.376829} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.904243] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Reconfigured VM instance instance-0000005a to attach disk [datastore1] e4fd9b9a-fc39-443c-9317-8dd286fdd2b0/e4fd9b9a-fc39-443c-9317-8dd286fdd2b0.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.904874] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c278b60-0c0c-45ff-b1d9-9d36428f3284 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.914010] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1031.914010] env[62952]: value = "task-1367508" [ 1031.914010] env[62952]: _type = "Task" [ 1031.914010] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.921228] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367508, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.950596] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.077979] env[62952]: DEBUG oslo_vmware.api [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367505, 'name': PowerOnVM_Task, 'duration_secs': 0.531239} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.080402] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.080646] env[62952]: INFO nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Took 7.89 seconds to spawn the instance on the hypervisor. [ 1032.080904] env[62952]: DEBUG nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.081746] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f788b68-4fcb-4e51-9a5e-2e02484560d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.234343] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a750511-c40f-4656-b8e8-8a2f063fd7b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.237130] env[62952]: DEBUG oslo_concurrency.lockutils [None req-bf0367ce-ab30-480a-801b-59312fe7aa8d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "1df00ddb-64fd-4ad3-be02-072a5aaad6f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.273s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.255431] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807570b8-60ea-4638-957d-777078e58c0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.296199] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367507, 'name': ReconfigVM_Task, 'duration_secs': 0.173953} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.296518] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1032.423310] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367508, 'name': Rename_Task, 'duration_secs': 0.142883} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.423574] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.423789] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90f82984-a354-4759-970d-0768ecbbebbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.430419] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1032.430419] env[62952]: value = "task-1367509" [ 1032.430419] env[62952]: _type = "Task" [ 1032.430419] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.438256] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.450076] env[62952]: DEBUG oslo_vmware.api [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367495, 'name': ReconfigVM_Task, 'duration_secs': 5.787225} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.450364] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.450578] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Reconfigured VM to detach interface {{(pid=62952) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1032.606070] env[62952]: INFO nova.compute.manager [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Took 17.17 seconds to build instance. [ 1032.707720] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.708014] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.768479] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1032.768479] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b64e0432-02bb-49b2-9412-d3096bcea43f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.774508] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1032.774508] env[62952]: value = "task-1367510" [ 1032.774508] env[62952]: _type = "Task" [ 1032.774508] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.786868] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367510, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.803328] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1032.803659] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1032.804273] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.804273] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1032.804360] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.804615] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1032.804729] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1032.804966] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1032.807029] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1032.807029] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1032.807029] env[62952]: DEBUG nova.virt.hardware [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1032.811311] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Reconfiguring VM instance instance-00000058 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1032.811967] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7cdae8f-f5e5-456b-8c8a-f12a9011bef4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.832123] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1032.832123] env[62952]: value = "task-1367511" [ 1032.832123] env[62952]: _type = "Task" [ 1032.832123] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.842616] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367511, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.953741] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367509, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.108589] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ca200ba-bf37-475b-964d-03daf4018c46 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.681s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.210522] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.286662] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367510, 'name': CreateSnapshot_Task, 'duration_secs': 0.45018} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.286957] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1033.287884] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076f9bdd-f9a8-44ff-a833-6dc3cac486f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.341667] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367511, 'name': ReconfigVM_Task, 'duration_secs': 0.2082} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.342096] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Reconfigured VM instance instance-00000058 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1033.342881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793fc19c-e8ff-44fc-ae52-d61cc69a5d5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.366964] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4/e0384d0d-6738-47f7-b14d-994cae2d1fe4.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.369195] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5692471c-b885-4483-b830-25408b051b39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.386151] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.386394] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.392702] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1033.392702] env[62952]: value = "task-1367512" [ 1033.392702] env[62952]: _type = "Task" [ 1033.392702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.433180] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367512, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.441248] env[62952]: DEBUG oslo_vmware.api [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367509, 'name': PowerOnVM_Task, 'duration_secs': 0.59217} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.441660] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.441880] env[62952]: INFO nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Took 15.33 seconds to spawn the instance on the hypervisor. [ 1033.442084] env[62952]: DEBUG nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.442889] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d99db07-1d47-4599-9f56-27b1c2547bb2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.733154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.733482] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.737347] env[62952]: INFO nova.compute.claims [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.787988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.788234] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquired lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.788426] env[62952]: DEBUG nova.network.neutron [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.806477] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1033.807148] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-53a7d165-dcb5-4084-a504-7b494d49f823 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.817951] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1033.817951] env[62952]: value = "task-1367513" [ 1033.817951] env[62952]: _type = "Task" [ 1033.817951] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.825944] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367513, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.888848] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.889120] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.889335] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.889523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.889786] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.891781] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.894344] env[62952]: INFO nova.compute.manager [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Terminating instance [ 1033.899252] env[62952]: DEBUG nova.compute.manager [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1033.899481] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.900294] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a81713-a5dc-4118-898b-5b796b70db57 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.908389] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.910721] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.910990] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5574b49b-2975-406b-b8f1-65575ba69239 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.916826] env[62952]: DEBUG oslo_vmware.api [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1033.916826] env[62952]: value = "task-1367514" [ 1033.916826] env[62952]: _type = "Task" [ 1033.916826] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.924744] env[62952]: DEBUG oslo_vmware.api [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.959121] env[62952]: INFO nova.compute.manager [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Took 20.25 seconds to build instance. [ 1034.245548] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.245801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.246014] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.246222] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.246393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.248450] env[62952]: INFO nova.compute.manager [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Terminating instance [ 1034.250261] env[62952]: DEBUG nova.compute.manager [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.250460] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1034.251302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27cfe4b-3b63-485c-9836-85d0274f0e5f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.260037] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.260522] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d08e3035-95e2-4495-929c-63e92b6b9153 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.267384] env[62952]: DEBUG oslo_vmware.api [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1034.267384] env[62952]: value = "task-1367515" [ 1034.267384] env[62952]: _type = "Task" [ 1034.267384] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.275396] env[62952]: DEBUG oslo_vmware.api [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.330596] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367513, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.413679] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367512, 'name': ReconfigVM_Task, 'duration_secs': 0.737966} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.413679] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Reconfigured VM instance instance-00000058 to attach disk [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4/e0384d0d-6738-47f7-b14d-994cae2d1fe4.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.413679] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.427050] env[62952]: DEBUG oslo_vmware.api [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367514, 'name': PowerOffVM_Task, 'duration_secs': 0.229815} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.427285] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.427409] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1034.428022] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-04a88572-8d42-4277-9291-ad948c8f4c30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.461761] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.463472] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e6e637cb-4b9c-449e-bcaf-c36ccc1f8952 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.763s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.489298] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1034.489502] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1034.489705] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore2] 2986a508-3f0b-4bad-b785-9e9f3ca7e39c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.489975] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fb1e398-e6fe-48b2-a62c-1447d7838c28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.499289] env[62952]: DEBUG oslo_vmware.api [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1034.499289] env[62952]: value = "task-1367517" [ 1034.499289] env[62952]: _type = "Task" [ 1034.499289] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.507681] env[62952]: DEBUG oslo_vmware.api [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367517, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.623520] env[62952]: INFO nova.network.neutron [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Port e1654a86-c311-45dd-9cc6-143cf350df90 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1034.623760] env[62952]: DEBUG nova.network.neutron [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [{"id": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "address": "fa:16:3e:ac:34:70", "network": {"id": "cab59c10-e233-4404-bbf5-40841bb12dbd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-407396040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d1f373267464f14ae2c833151821973", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a39f7b2-d4", "ovs_interfaceid": "3a39f7b2-d402-4ead-8172-34ae1a9c4387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.777549] env[62952]: DEBUG oslo_vmware.api [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367515, 'name': PowerOffVM_Task, 'duration_secs': 0.222512} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.777865] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.778052] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1034.778295] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ec65012-a80b-4d0f-ad32-d645610070fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.828404] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367513, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.862999] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1034.863319] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1034.863560] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleting the datastore file [datastore2] 303c9fe7-0d89-4a1a-adcc-6f787486df61 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.863875] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78560c0d-b727-4aaa-ae62-62dac52f5a11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.871412] env[62952]: DEBUG oslo_vmware.api [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1034.871412] env[62952]: value = "task-1367519" [ 1034.871412] env[62952]: _type = "Task" [ 1034.871412] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.879667] env[62952]: DEBUG oslo_vmware.api [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.925317] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e98995f-e67d-4c7b-8ae8-dd47b7e77a5f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.946579] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f218ef-6a9e-47b1-89c7-ac9f8b064f99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.964140] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.010450] env[62952]: DEBUG oslo_vmware.api [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367517, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221262} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.011555] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.011706] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1035.011898] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1035.012183] env[62952]: INFO nova.compute.manager [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1035.012409] env[62952]: DEBUG oslo.service.loopingcall [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.016205] env[62952]: DEBUG nova.compute.manager [-] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.016394] env[62952]: DEBUG nova.network.neutron [-] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1035.125336] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ef426d-92c8-422f-b0de-027594dda935 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.130440] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Releasing lock "refresh_cache-303c9fe7-0d89-4a1a-adcc-6f787486df61" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.137217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4388b860-8489-427f-a84e-5f3bdba7f30c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.166855] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf34e0c-efc3-4848-abeb-3c752a186db3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.174026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad3e0a1-3a57-44c0-9b7e-fb63e659ad11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.187062] env[62952]: DEBUG nova.compute.provider_tree [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.288202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.288202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.288202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.288202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.288202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.292673] env[62952]: INFO nova.compute.manager [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Terminating instance [ 1035.300224] env[62952]: DEBUG nova.compute.manager [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.300438] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.301303] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51daf5f-6d8a-4fb4-ad67-5911abc7102d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.311466] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.311466] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2baa528-da53-479c-9f5a-5896d4164289 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.319855] env[62952]: DEBUG oslo_vmware.api [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1035.319855] env[62952]: value = "task-1367520" [ 1035.319855] env[62952]: _type = "Task" [ 1035.319855] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.335357] env[62952]: DEBUG oslo_vmware.api [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.340486] env[62952]: DEBUG nova.compute.manager [req-7adfc8b5-c23b-47df-9c40-b7715a3accc1 req-3ad3040f-0752-4de7-8ee8-d4543709e2a8 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Received event network-vif-deleted-8514c98b-afbf-4e2a-bccf-d33990628a2e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.340690] env[62952]: INFO nova.compute.manager [req-7adfc8b5-c23b-47df-9c40-b7715a3accc1 req-3ad3040f-0752-4de7-8ee8-d4543709e2a8 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Neutron deleted interface 8514c98b-afbf-4e2a-bccf-d33990628a2e; detaching it from the instance and deleting it from the info cache [ 1035.340894] env[62952]: DEBUG nova.network.neutron [req-7adfc8b5-c23b-47df-9c40-b7715a3accc1 req-3ad3040f-0752-4de7-8ee8-d4543709e2a8 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.342197] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367513, 'name': CloneVM_Task, 'duration_secs': 1.211976} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.343088] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Created linked-clone VM from snapshot [ 1035.343939] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7192ed-beb2-4f23-ab27-9a5a42052c21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.351681] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Uploading image 2e8bcbde-a25a-44b1-80bc-4205a778724e {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1035.376255] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1035.376255] env[62952]: value = "vm-290964" [ 1035.376255] env[62952]: _type = "VirtualMachine" [ 1035.376255] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1035.376956] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8fc0c443-40ac-439a-baa4-013aa724f743 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.383537] env[62952]: DEBUG oslo_vmware.api [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.384689] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.384908] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1035.385155] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1035.385344] env[62952]: INFO nova.compute.manager [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1035.385592] env[62952]: DEBUG oslo.service.loopingcall [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.385911] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease: (returnval){ [ 1035.385911] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a14266-a17e-4af1-226b-44a83bf04bdf" [ 1035.385911] env[62952]: _type = "HttpNfcLease" [ 1035.385911] env[62952]: } obtained for exporting VM: (result){ [ 1035.385911] env[62952]: value = "vm-290964" [ 1035.385911] env[62952]: _type = "VirtualMachine" [ 1035.385911] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1035.386227] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the lease: (returnval){ [ 1035.386227] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a14266-a17e-4af1-226b-44a83bf04bdf" [ 1035.386227] env[62952]: _type = "HttpNfcLease" [ 1035.386227] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1035.386409] env[62952]: DEBUG nova.compute.manager [-] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.386519] env[62952]: DEBUG nova.network.neutron [-] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1035.396518] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1035.396518] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a14266-a17e-4af1-226b-44a83bf04bdf" [ 1035.396518] env[62952]: _type = "HttpNfcLease" [ 1035.396518] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1035.396797] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1035.396797] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a14266-a17e-4af1-226b-44a83bf04bdf" [ 1035.396797] env[62952]: _type = "HttpNfcLease" [ 1035.396797] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1035.397551] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e0c08d-b24b-42e5-805e-194d35778609 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.404597] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52605e38-777f-8c91-6350-dbb2d54b8591/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1035.404775] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52605e38-777f-8c91-6350-dbb2d54b8591/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1035.495239] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8c5a555b-0395-44fc-a2aa-74a765b428b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.537387] env[62952]: DEBUG nova.network.neutron [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Port 3f88a864-36dc-4fc8-8073-977444f51ed8 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1035.634363] env[62952]: DEBUG oslo_concurrency.lockutils [None req-51a79d48-db3a-4eff-9f9e-fc68c8a324a6 tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "interface-303c9fe7-0d89-4a1a-adcc-6f787486df61-e1654a86-c311-45dd-9cc6-143cf350df90" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.769s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.693431] env[62952]: DEBUG nova.scheduler.client.report [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.805333] env[62952]: DEBUG nova.network.neutron [-] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.830629] env[62952]: DEBUG oslo_vmware.api [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367520, 'name': PowerOffVM_Task, 'duration_secs': 0.333265} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.831273] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.831464] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.831782] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc50abdb-73a8-4dfa-aa72-6dd8f14c4902 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.843962] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-915d100f-bf80-4778-a245-0e305d19f664 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.853758] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d45a206-cd64-4be1-a764-3def4cfbd2c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.883967] env[62952]: DEBUG nova.compute.manager [req-7adfc8b5-c23b-47df-9c40-b7715a3accc1 req-3ad3040f-0752-4de7-8ee8-d4543709e2a8 service nova] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Detach interface failed, port_id=8514c98b-afbf-4e2a-bccf-d33990628a2e, reason: Instance 2986a508-3f0b-4bad-b785-9e9f3ca7e39c could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1035.987788] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.988273] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.988549] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleting the datastore file [datastore1] e4fd9b9a-fc39-443c-9317-8dd286fdd2b0 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.989415] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4cb1b738-ec5c-4c72-b18c-5f4abedd640e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.995869] env[62952]: DEBUG oslo_vmware.api [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1035.995869] env[62952]: value = "task-1367523" [ 1035.995869] env[62952]: _type = "Task" [ 1035.995869] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.004771] env[62952]: DEBUG oslo_vmware.api [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367523, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.199010] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.203031] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.204964] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.743s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.209993] env[62952]: INFO nova.compute.claims [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.308632] env[62952]: INFO nova.compute.manager [-] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Took 1.29 seconds to deallocate network for instance. [ 1036.509295] env[62952]: DEBUG oslo_vmware.api [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367523, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.509295] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.509295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.509295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.509295] env[62952]: INFO nova.compute.manager [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1036.509736] env[62952]: DEBUG oslo.service.loopingcall [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.509736] env[62952]: DEBUG nova.compute.manager [-] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.509814] env[62952]: DEBUG nova.network.neutron [-] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.561853] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.562260] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.562621] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.711900] env[62952]: DEBUG nova.compute.utils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.718361] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1036.720831] env[62952]: DEBUG nova.network.neutron [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1036.786193] env[62952]: DEBUG nova.policy [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '913eacf6a94845b189e0bd5df408138a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8c2a296e8fe4805b9bb761f91a407f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.819027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.014485] env[62952]: DEBUG nova.network.neutron [-] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.083600] env[62952]: DEBUG nova.network.neutron [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Successfully created port: be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.219783] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.236176] env[62952]: DEBUG nova.network.neutron [-] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.383060] env[62952]: DEBUG nova.compute.manager [req-9d172826-2fcc-48cd-9507-652d0bc1b887 req-dcf22fcb-5f69-496d-8b96-ca318dcad1b8 service nova] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Received event network-vif-deleted-3a39f7b2-d402-4ead-8172-34ae1a9c4387 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.383323] env[62952]: DEBUG nova.compute.manager [req-9d172826-2fcc-48cd-9507-652d0bc1b887 req-dcf22fcb-5f69-496d-8b96-ca318dcad1b8 service nova] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Received event network-vif-deleted-dfb9de49-97fd-4f78-8770-4ac0e79bd8ed {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.475790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad835d5-4dfd-4cb9-82e4-f6800d3314af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.484023] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e4b2e8-bfcc-4f2d-8fe1-585ee3fae2b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.518583] env[62952]: INFO nova.compute.manager [-] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Took 2.13 seconds to deallocate network for instance. [ 1037.519508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96659ad-7eda-44d9-8ff1-963a5a99479b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.530470] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e14f18-38ba-4f1d-836f-d76b59e86523 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.545119] env[62952]: DEBUG nova.compute.provider_tree [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.626356] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.626571] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.626756] env[62952]: DEBUG nova.network.neutron [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.738390] env[62952]: INFO nova.compute.manager [-] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Took 1.23 seconds to deallocate network for instance. [ 1037.847897] env[62952]: DEBUG oslo_concurrency.lockutils [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.848254] env[62952]: DEBUG oslo_concurrency.lockutils [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.031222] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.048527] env[62952]: DEBUG nova.scheduler.client.report [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.233372] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.244551] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.260448] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.260713] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.260880] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.261116] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.261291] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.261446] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.261657] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.261819] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.261993] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.262217] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.262386] env[62952]: DEBUG nova.virt.hardware [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.263231] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8892ed74-942c-461a-ad24-314ef851d6ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.273490] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cf10fb-2fc1-49cf-897b-5b96c1ae7928 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.351233] env[62952]: INFO nova.compute.manager [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Detaching volume 5b018b5e-81cf-4820-85bd-d3015cbda494 [ 1038.384754] env[62952]: DEBUG nova.network.neutron [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance_info_cache with network_info: [{"id": "3f88a864-36dc-4fc8-8073-977444f51ed8", "address": "fa:16:3e:70:82:3d", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f88a864-36", "ovs_interfaceid": "3f88a864-36dc-4fc8-8073-977444f51ed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.395081] env[62952]: INFO nova.virt.block_device [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Attempting to driver detach volume 5b018b5e-81cf-4820-85bd-d3015cbda494 from mountpoint /dev/sdb [ 1038.395265] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1038.395609] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290942', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'name': 'volume-5b018b5e-81cf-4820-85bd-d3015cbda494', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b', 'attached_at': '', 'detached_at': '', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'serial': '5b018b5e-81cf-4820-85bd-d3015cbda494'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1038.397064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20c0359-f930-4bc7-82ac-efc02e15e0a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.420085] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00da79a6-009d-4548-bf01-e29b6e9824e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.427059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f5074f-2f96-458d-8ac3-70a18823e75e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.431311] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.431577] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.449123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf722a2f-4be9-41af-8399-c398b6679e6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.463558] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] The volume has not been displaced from its original location: [datastore1] volume-5b018b5e-81cf-4820-85bd-d3015cbda494/volume-5b018b5e-81cf-4820-85bd-d3015cbda494.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1038.468766] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1038.469931] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-677a7586-5ed7-4145-9ee2-18c929927758 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.488569] env[62952]: DEBUG oslo_vmware.api [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1038.488569] env[62952]: value = "task-1367524" [ 1038.488569] env[62952]: _type = "Task" [ 1038.488569] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.497039] env[62952]: DEBUG oslo_vmware.api [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.553555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.554225] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.557283] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.738s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.557576] env[62952]: DEBUG nova.objects.instance [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid 2986a508-3f0b-4bad-b785-9e9f3ca7e39c {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.737451] env[62952]: DEBUG nova.network.neutron [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Successfully updated port: be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.886688] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.954501] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.954501] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1038.998781] env[62952]: DEBUG oslo_vmware.api [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367524, 'name': ReconfigVM_Task, 'duration_secs': 0.244195} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.999465] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1039.004526] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03043ea8-63bf-413b-852e-7c293272b135 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.021752] env[62952]: DEBUG oslo_vmware.api [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1039.021752] env[62952]: value = "task-1367525" [ 1039.021752] env[62952]: _type = "Task" [ 1039.021752] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.030775] env[62952]: DEBUG oslo_vmware.api [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.060565] env[62952]: DEBUG nova.compute.utils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.062560] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1039.062880] env[62952]: DEBUG nova.network.neutron [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1039.118899] env[62952]: DEBUG nova.policy [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f24a30e1160a43d5be6cde39b8f79edd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '094ca56818604bee9c1f75d1493ab3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1039.240930] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.241808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.241808] env[62952]: DEBUG nova.network.neutron [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.289612] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54126367-39f9-4e39-a889-d66b29cb16d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.297093] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356021f0-92a5-49a0-8c69-a20defb22c21 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.329746] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3db601-7d33-40be-8972-79c09aee0232 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.337370] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd4679f-cbed-4a7a-92d0-9ac4b72a4395 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.352794] env[62952]: DEBUG nova.compute.provider_tree [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.399169] env[62952]: DEBUG nova.network.neutron [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Successfully created port: 9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.410206] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fae7f8c-466c-4bad-a2a3-9351ba36180e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.415586] env[62952]: DEBUG nova.compute.manager [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Received event network-vif-plugged-be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.415842] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.416148] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.416301] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.416490] env[62952]: DEBUG nova.compute.manager [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] No waiting events found dispatching network-vif-plugged-be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.416685] env[62952]: WARNING nova.compute.manager [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Received unexpected event network-vif-plugged-be83803d-9485-4459-882a-5dc9383e0f62 for instance with vm_state building and task_state spawning. [ 1039.416882] env[62952]: DEBUG nova.compute.manager [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Received event network-changed-be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.417079] env[62952]: DEBUG nova.compute.manager [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Refreshing instance network info cache due to event network-changed-be83803d-9485-4459-882a-5dc9383e0f62. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.417286] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] Acquiring lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.442743] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06aa3f22-55b5-44a9-998b-7a7eac890afd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.451240] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.536638] env[62952]: DEBUG oslo_vmware.api [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367525, 'name': ReconfigVM_Task, 'duration_secs': 0.191005} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.536638] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290942', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'name': 'volume-5b018b5e-81cf-4820-85bd-d3015cbda494', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b', 'attached_at': '', 'detached_at': '', 'volume_id': '5b018b5e-81cf-4820-85bd-d3015cbda494', 'serial': '5b018b5e-81cf-4820-85bd-d3015cbda494'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1039.570055] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.774378] env[62952]: DEBUG nova.network.neutron [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1039.855625] env[62952]: DEBUG nova.scheduler.client.report [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.938613] env[62952]: DEBUG nova.network.neutron [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating instance_info_cache with network_info: [{"id": "be83803d-9485-4459-882a-5dc9383e0f62", "address": "fa:16:3e:0b:bc:d7", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe83803d-94", "ovs_interfaceid": "be83803d-9485-4459-882a-5dc9383e0f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.961953] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.963660] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fae3e351-7c80-4add-a1f3-3260da878d95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.974748] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1039.974748] env[62952]: value = "task-1367526" [ 1039.974748] env[62952]: _type = "Task" [ 1039.974748] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.983804] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.078323] env[62952]: DEBUG nova.objects.instance [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.360668] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.362929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.332s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.363228] env[62952]: DEBUG nova.objects.instance [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lazy-loading 'resources' on Instance uuid 303c9fe7-0d89-4a1a-adcc-6f787486df61 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.385176] env[62952]: INFO nova.scheduler.client.report [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance 2986a508-3f0b-4bad-b785-9e9f3ca7e39c [ 1040.441755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.442129] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance network_info: |[{"id": "be83803d-9485-4459-882a-5dc9383e0f62", "address": "fa:16:3e:0b:bc:d7", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe83803d-94", "ovs_interfaceid": "be83803d-9485-4459-882a-5dc9383e0f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1040.442476] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] Acquired lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.442670] env[62952]: DEBUG nova.network.neutron [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Refreshing network info cache for port be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.443947] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:bc:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be83803d-9485-4459-882a-5dc9383e0f62', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.451717] env[62952]: DEBUG oslo.service.loopingcall [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.452664] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1040.452910] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67f4f28f-e5c4-4fb8-ae7c-00f354fdeb28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.480936] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.480936] env[62952]: value = "task-1367527" [ 1040.480936] env[62952]: _type = "Task" [ 1040.480936] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.487679] env[62952]: DEBUG oslo_vmware.api [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367526, 'name': PowerOnVM_Task, 'duration_secs': 0.38794} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.488445] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.488694] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd0db84-9b0f-43a4-9ac8-ba0f95667021 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance 'e0384d0d-6738-47f7-b14d-994cae2d1fe4' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1040.495945] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367527, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.500955] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.501117] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.501318] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1040.582179] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.614193] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.614503] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.614689] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.614887] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.615077] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.615237] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.615450] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.615628] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.615821] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.616033] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.616232] env[62952]: DEBUG nova.virt.hardware [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.617154] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688c2b22-0feb-4a69-8e57-b89e7af36379 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.627305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddda8c1-325b-41cf-9b71-d4edc1c62176 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.893280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a7218bbd-8e74-463e-8607-986ae35357ac tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2986a508-3f0b-4bad-b785-9e9f3ca7e39c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.004s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.971765] env[62952]: DEBUG nova.network.neutron [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Successfully updated port: 9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.994217] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367527, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.059282] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.093197] env[62952]: DEBUG oslo_concurrency.lockutils [None req-98012544-5440-4c18-94f8-f156a7da48d0 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.245s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.094303] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.035s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.094418] env[62952]: DEBUG nova.compute.manager [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.095535] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a3d359-6ab1-4fbc-a8e3-eeea87bb905d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.104772] env[62952]: DEBUG nova.compute.manager [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1041.105281] env[62952]: DEBUG nova.objects.instance [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.109878] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96caa06c-75d9-429d-80fd-59661df1d204 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.122547] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40eac60-7e97-4179-bbf5-a2b45bdf2d91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.159137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c588c68a-9399-4bae-ae0f-d725b7592387 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.169419] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451ef1f6-4076-497b-9bb9-5f72385ef3b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.184153] env[62952]: DEBUG nova.compute.provider_tree [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.292868] env[62952]: DEBUG nova.network.neutron [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updated VIF entry in instance network info cache for port be83803d-9485-4459-882a-5dc9383e0f62. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1041.293347] env[62952]: DEBUG nova.network.neutron [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating instance_info_cache with network_info: [{"id": "be83803d-9485-4459-882a-5dc9383e0f62", "address": "fa:16:3e:0b:bc:d7", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe83803d-94", "ovs_interfaceid": "be83803d-9485-4459-882a-5dc9383e0f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.473524] env[62952]: DEBUG nova.compute.manager [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Received event network-vif-plugged-9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.477372] env[62952]: DEBUG oslo_concurrency.lockutils [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] Acquiring lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.477372] env[62952]: DEBUG oslo_concurrency.lockutils [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.477372] env[62952]: DEBUG oslo_concurrency.lockutils [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.477372] env[62952]: DEBUG nova.compute.manager [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] No waiting events found dispatching network-vif-plugged-9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1041.477372] env[62952]: WARNING nova.compute.manager [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Received unexpected event network-vif-plugged-9cf16459-ba22-47ad-82f5-735a585a9e9e for instance with vm_state building and task_state spawning. [ 1041.477372] env[62952]: DEBUG nova.compute.manager [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Received event network-changed-9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.477372] env[62952]: DEBUG nova.compute.manager [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Refreshing instance network info cache due to event network-changed-9cf16459-ba22-47ad-82f5-735a585a9e9e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1041.479238] env[62952]: DEBUG oslo_concurrency.lockutils [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] Acquiring lock "refresh_cache-faa3b525-bd20-4e54-ba4c-60d65bf601aa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.479406] env[62952]: DEBUG oslo_concurrency.lockutils [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] Acquired lock "refresh_cache-faa3b525-bd20-4e54-ba4c-60d65bf601aa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.479584] env[62952]: DEBUG nova.network.neutron [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Refreshing network info cache for port 9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1041.481183] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-faa3b525-bd20-4e54-ba4c-60d65bf601aa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.497219] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367527, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.615232] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.615542] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f726c173-c863-4f08-9370-302718c65e9b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.623908] env[62952]: DEBUG oslo_vmware.api [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1041.623908] env[62952]: value = "task-1367528" [ 1041.623908] env[62952]: _type = "Task" [ 1041.623908] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.632666] env[62952]: DEBUG oslo_vmware.api [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.686131] env[62952]: DEBUG nova.scheduler.client.report [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.788043] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.796934] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb1d8f4f-19cb-4ff3-85f0-14778f320151 req-221e99c4-16bc-4e74-b6ff-72d77864bf70 service nova] Releasing lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.998852] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367527, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.025958] env[62952]: DEBUG nova.network.neutron [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1042.081267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.081889] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.135840] env[62952]: DEBUG oslo_vmware.api [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.176319] env[62952]: DEBUG nova.network.neutron [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.191349] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.194837] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.949s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.194837] env[62952]: DEBUG nova.objects.instance [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lazy-loading 'resources' on Instance uuid e4fd9b9a-fc39-443c-9317-8dd286fdd2b0 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.216602] env[62952]: INFO nova.scheduler.client.report [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleted allocations for instance 303c9fe7-0d89-4a1a-adcc-6f787486df61 [ 1042.290902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.291244] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1042.291474] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.291808] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.292082] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.292316] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.292600] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.292781] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.292940] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1042.293218] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.495947] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367527, 'name': CreateVM_Task, 'duration_secs': 1.55976} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.496151] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.496856] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.497044] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.497426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.497715] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bb546d6-6e8c-44a0-b841-bd444f353090 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.503146] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1042.503146] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52043b0a-d153-eeda-9a29-9c2e4069751c" [ 1042.503146] env[62952]: _type = "Task" [ 1042.503146] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.514221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.514464] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.514653] env[62952]: DEBUG nova.compute.manager [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Going to confirm migration 1 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1042.515958] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52043b0a-d153-eeda-9a29-9c2e4069751c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.584474] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1042.635725] env[62952]: DEBUG oslo_vmware.api [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367528, 'name': PowerOffVM_Task, 'duration_secs': 0.598987} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.636045] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.636222] env[62952]: DEBUG nova.compute.manager [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.637063] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b059ba4b-e51f-457a-975d-56209a23c6ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.679437] env[62952]: DEBUG oslo_concurrency.lockutils [req-25ebb313-6577-44f3-a472-cbaa0c9aefd4 req-a3c0a2c1-c086-48d5-be55-815db36a4989 service nova] Releasing lock "refresh_cache-faa3b525-bd20-4e54-ba4c-60d65bf601aa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.680035] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-faa3b525-bd20-4e54-ba4c-60d65bf601aa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.680194] env[62952]: DEBUG nova.network.neutron [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.725738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ed77d114-b45d-492a-b60b-c6143894b54f tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "303c9fe7-0d89-4a1a-adcc-6f787486df61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.480s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.797156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.881283] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b5cf17-0a42-4fcf-97fd-617aab12f6b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.889338] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36de239-5fa7-4fa0-96af-0ce3c3f4e443 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.920790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7a6504-0188-4849-9e85-423bd4c6d041 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.930955] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe29262-442e-4d49-ae41-1fe71d4d1560 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.946434] env[62952]: DEBUG nova.compute.provider_tree [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.955087] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52605e38-777f-8c91-6350-dbb2d54b8591/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1042.955911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d52468-30cf-40a2-b331-8dc68be5d8ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.962827] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52605e38-777f-8c91-6350-dbb2d54b8591/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1042.962997] env[62952]: ERROR oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52605e38-777f-8c91-6350-dbb2d54b8591/disk-0.vmdk due to incomplete transfer. [ 1042.963230] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2a930c47-5e5e-4c7a-94eb-4996fd835d7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.971743] env[62952]: DEBUG oslo_vmware.rw_handles [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52605e38-777f-8c91-6350-dbb2d54b8591/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1042.971987] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Uploaded image 2e8bcbde-a25a-44b1-80bc-4205a778724e to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1042.974554] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1042.974823] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4ac63653-8792-4a7d-9150-98e3f2a3685d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.983179] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1042.983179] env[62952]: value = "task-1367529" [ 1042.983179] env[62952]: _type = "Task" [ 1042.983179] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.991398] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367529, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.019984] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52043b0a-d153-eeda-9a29-9c2e4069751c, 'name': SearchDatastore_Task, 'duration_secs': 0.011184} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.022457] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.022691] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.022930] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.023121] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.023320] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.024028] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ee628dd-cca4-47e2-97ed-a3196be2a5e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.035590] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.036398] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.036547] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e7a0c0-919b-469b-99de-79154e882e3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.043032] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1043.043032] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5203063a-1696-c813-99bc-a4216b6d6e68" [ 1043.043032] env[62952]: _type = "Task" [ 1043.043032] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.051841] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5203063a-1696-c813-99bc-a4216b6d6e68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.078924] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.079125] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.079312] env[62952]: DEBUG nova.network.neutron [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1043.079501] env[62952]: DEBUG nova.objects.instance [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'info_cache' on Instance uuid e0384d0d-6738-47f7-b14d-994cae2d1fe4 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.107348] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.151109] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2526b56d-37bb-4058-88db-4193eb8b2759 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.056s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.208573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.208835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.209059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.209245] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.209412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.211606] env[62952]: INFO nova.compute.manager [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Terminating instance [ 1043.213548] env[62952]: DEBUG nova.compute.manager [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1043.213756] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.214603] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8072e1-1a2a-4973-bc26-c0537c2baf7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.224222] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.224222] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-008becba-7350-42d6-9fde-d7664103efd3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.225941] env[62952]: DEBUG nova.network.neutron [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1043.232819] env[62952]: DEBUG oslo_vmware.api [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1043.232819] env[62952]: value = "task-1367530" [ 1043.232819] env[62952]: _type = "Task" [ 1043.232819] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.240988] env[62952]: DEBUG oslo_vmware.api [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367530, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.401666] env[62952]: DEBUG nova.network.neutron [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Updating instance_info_cache with network_info: [{"id": "9cf16459-ba22-47ad-82f5-735a585a9e9e", "address": "fa:16:3e:f7:81:10", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cf16459-ba", "ovs_interfaceid": "9cf16459-ba22-47ad-82f5-735a585a9e9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.445116] env[62952]: DEBUG oslo_concurrency.lockutils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.445814] env[62952]: DEBUG oslo_concurrency.lockutils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.450366] env[62952]: DEBUG nova.scheduler.client.report [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.492870] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367529, 'name': Destroy_Task, 'duration_secs': 0.391537} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.493185] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroyed the VM [ 1043.493431] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1043.493683] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-97952a67-ca59-423c-b314-c76ed398b723 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.501706] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1043.501706] env[62952]: value = "task-1367531" [ 1043.501706] env[62952]: _type = "Task" [ 1043.501706] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.512286] env[62952]: DEBUG nova.objects.instance [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.513529] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367531, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.553646] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5203063a-1696-c813-99bc-a4216b6d6e68, 'name': SearchDatastore_Task, 'duration_secs': 0.010159} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.554499] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3386fea9-24fb-4a48-adaf-85413c237d5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.560256] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1043.560256] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]529414c0-d0d8-caba-f12a-df5cdf5bf44a" [ 1043.560256] env[62952]: _type = "Task" [ 1043.560256] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.568100] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529414c0-d0d8-caba-f12a-df5cdf5bf44a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.743432] env[62952]: DEBUG oslo_vmware.api [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367530, 'name': PowerOffVM_Task, 'duration_secs': 0.206405} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.743794] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.743917] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.744156] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36f3026d-18cf-4f41-905e-a966f329f4d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.820831] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1043.821131] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1043.821353] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleting the datastore file [datastore2] ff189c9e-9e7c-4217-9c65-0f821393870a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.821724] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92fd0df5-4f5a-4f2b-a0c6-c274cdd0b5ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.829055] env[62952]: DEBUG oslo_vmware.api [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for the task: (returnval){ [ 1043.829055] env[62952]: value = "task-1367533" [ 1043.829055] env[62952]: _type = "Task" [ 1043.829055] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.838337] env[62952]: DEBUG oslo_vmware.api [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.904801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-faa3b525-bd20-4e54-ba4c-60d65bf601aa" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.905230] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Instance network_info: |[{"id": "9cf16459-ba22-47ad-82f5-735a585a9e9e", "address": "fa:16:3e:f7:81:10", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cf16459-ba", "ovs_interfaceid": "9cf16459-ba22-47ad-82f5-735a585a9e9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1043.905722] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:81:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97113f46-d648-4613-b233-069acba18198', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cf16459-ba22-47ad-82f5-735a585a9e9e', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1043.913208] env[62952]: DEBUG oslo.service.loopingcall [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.913445] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1043.913675] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4fba87be-cdc5-4f58-b402-a9719bbe183e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.935400] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.935400] env[62952]: value = "task-1367534" [ 1043.935400] env[62952]: _type = "Task" [ 1043.935400] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.943335] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367534, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.948424] env[62952]: DEBUG nova.compute.utils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.954755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.957914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.160s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.957914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.957914] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1043.957914] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.851s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.959474] env[62952]: INFO nova.compute.claims [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.962474] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e87a2f-0826-41ba-9e1a-46b8a6e9e00b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.973895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a851f3-e76d-4845-9d7c-4a2c2045f45f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.979152] env[62952]: INFO nova.scheduler.client.report [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted allocations for instance e4fd9b9a-fc39-443c-9317-8dd286fdd2b0 [ 1043.993348] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be74baf-a260-47c8-8a2c-77ca51c524fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.002615] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe80896-b73c-4151-acf1-e62553e0c3ba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.015778] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367531, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.044489] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179933MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1044.044712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.045285] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.045463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.045631] env[62952]: DEBUG nova.network.neutron [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.045812] env[62952]: DEBUG nova.objects.instance [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'info_cache' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.071233] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529414c0-d0d8-caba-f12a-df5cdf5bf44a, 'name': SearchDatastore_Task, 'duration_secs': 0.264154} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.071534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.071837] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.072137] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26cee18f-1345-4119-9cf8-6d2c90a85952 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.080444] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1044.080444] env[62952]: value = "task-1367535" [ 1044.080444] env[62952]: _type = "Task" [ 1044.080444] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.089613] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367535, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.298468] env[62952]: DEBUG nova.network.neutron [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance_info_cache with network_info: [{"id": "3f88a864-36dc-4fc8-8073-977444f51ed8", "address": "fa:16:3e:70:82:3d", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f88a864-36", "ovs_interfaceid": "3f88a864-36dc-4fc8-8073-977444f51ed8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.341875] env[62952]: DEBUG oslo_vmware.api [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Task: {'id': task-1367533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16918} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.343073] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.343337] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.343675] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.344013] env[62952]: INFO nova.compute.manager [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1044.344426] env[62952]: DEBUG oslo.service.loopingcall [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.344752] env[62952]: DEBUG nova.compute.manager [-] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.344917] env[62952]: DEBUG nova.network.neutron [-] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1044.448017] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367534, 'name': CreateVM_Task, 'duration_secs': 0.397519} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.448229] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.448905] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.449083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.449417] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.449693] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5102085b-578e-499d-9836-81ee721440c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.451855] env[62952]: DEBUG oslo_concurrency.lockutils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.456562] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1044.456562] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]520f43d3-99f6-da6e-a76d-58777076c395" [ 1044.456562] env[62952]: _type = "Task" [ 1044.456562] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.465938] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520f43d3-99f6-da6e-a76d-58777076c395, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.486633] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0e846ea6-b332-4db7-9601-068e282d7738 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "e4fd9b9a-fc39-443c-9317-8dd286fdd2b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.199s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.522976] env[62952]: DEBUG oslo_vmware.api [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367531, 'name': RemoveSnapshot_Task, 'duration_secs': 0.527703} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.523479] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1044.523863] env[62952]: INFO nova.compute.manager [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 12.29 seconds to snapshot the instance on the hypervisor. [ 1044.549419] env[62952]: DEBUG nova.objects.base [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Object Instance<88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b> lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1044.591097] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367535, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.591393] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.591663] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.591928] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b34a0f30-ff6d-4fdc-8a7b-3587f99c49cf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.598866] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1044.598866] env[62952]: value = "task-1367536" [ 1044.598866] env[62952]: _type = "Task" [ 1044.598866] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.606859] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367536, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.659508] env[62952]: DEBUG nova.compute.manager [req-11870bd9-23cf-4802-93f1-17f48d28e34d req-d43851c8-ce6f-4b79-adfa-e10fac669b0d service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Received event network-vif-deleted-48403971-25e5-4da0-a350-bd47ba090afc {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.659754] env[62952]: INFO nova.compute.manager [req-11870bd9-23cf-4802-93f1-17f48d28e34d req-d43851c8-ce6f-4b79-adfa-e10fac669b0d service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Neutron deleted interface 48403971-25e5-4da0-a350-bd47ba090afc; detaching it from the instance and deleting it from the info cache [ 1044.659910] env[62952]: DEBUG nova.network.neutron [req-11870bd9-23cf-4802-93f1-17f48d28e34d req-d43851c8-ce6f-4b79-adfa-e10fac669b0d service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.802136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-e0384d0d-6738-47f7-b14d-994cae2d1fe4" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.802471] env[62952]: DEBUG nova.objects.instance [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'migration_context' on Instance uuid e0384d0d-6738-47f7-b14d-994cae2d1fe4 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.967633] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520f43d3-99f6-da6e-a76d-58777076c395, 'name': SearchDatastore_Task, 'duration_secs': 0.05379} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.968028] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.968260] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.968503] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.968654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.968835] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.971625] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0d46325-b875-447a-bb64-39cebb3ef3de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.982617] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.982889] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.983655] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ce4555b-d576-4046-8f3f-e07befc8e509 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.990164] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1044.990164] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5200a1a7-8ec1-6482-833b-799e3f367d13" [ 1044.990164] env[62952]: _type = "Task" [ 1044.990164] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.005207] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5200a1a7-8ec1-6482-833b-799e3f367d13, 'name': SearchDatastore_Task, 'duration_secs': 0.010809} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.008710] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f8d3ee1-d765-4a1c-8de5-47fdb51ca8b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.014952] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1045.014952] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524da1e5-5a9a-ae61-6895-07548b01fc29" [ 1045.014952] env[62952]: _type = "Task" [ 1045.014952] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.032428] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524da1e5-5a9a-ae61-6895-07548b01fc29, 'name': SearchDatastore_Task, 'duration_secs': 0.010312} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.035152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.035413] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] faa3b525-bd20-4e54-ba4c-60d65bf601aa/faa3b525-bd20-4e54-ba4c-60d65bf601aa.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.037871] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6146b9cb-3bdd-4d66-8499-65ff9231f79a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.046807] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1045.046807] env[62952]: value = "task-1367537" [ 1045.046807] env[62952]: _type = "Task" [ 1045.046807] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.058292] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367537, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.069664] env[62952]: DEBUG nova.compute.manager [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Found 3 images (rotation: 2) {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1045.069886] env[62952]: DEBUG nova.compute.manager [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Rotating out 1 backups {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1045.070068] env[62952]: DEBUG nova.compute.manager [None req-9056cc33-4427-41c1-9958-287e80260ff2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleting image 497ca4b0-de44-46b4-b59e-8ecf1e60917f {{(pid=62952) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1045.078907] env[62952]: DEBUG nova.network.neutron [-] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.108800] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367536, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069899} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.111475] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.112785] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f014673c-bd77-436a-8395-34f646bcd686 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.140397] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.143248] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4888cd29-3552-4c49-8b39-6e8319c7958f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.163732] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a564407-eb52-43a8-907b-1a1bacb6b0f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.168268] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1045.168268] env[62952]: value = "task-1367538" [ 1045.168268] env[62952]: _type = "Task" [ 1045.168268] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.179058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff478e6-750a-42ff-8440-20c1a0488090 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.197795] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.220173] env[62952]: DEBUG nova.compute.manager [req-11870bd9-23cf-4802-93f1-17f48d28e34d req-d43851c8-ce6f-4b79-adfa-e10fac669b0d service nova] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Detach interface failed, port_id=48403971-25e5-4da0-a350-bd47ba090afc, reason: Instance ff189c9e-9e7c-4217-9c65-0f821393870a could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1045.222089] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e28a002-15ca-4946-bca6-eb8463d25932 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.229224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ccac33-b9e1-4ac1-837e-b23ccf299fb2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.259753] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211d337f-732b-4dfc-b6b8-7cb6f0236fff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.267939] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62368ab-99ef-4d9f-b1cd-d4e110c2fc38 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.284792] env[62952]: DEBUG nova.compute.provider_tree [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.305945] env[62952]: DEBUG nova.objects.base [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1045.307162] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2202ada-7506-4975-b727-eb89838db346 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.330585] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2326ab04-f9f6-49b9-b90e-35dde1a614b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.332955] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.333205] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.333410] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.333596] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.333767] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.336096] env[62952]: INFO nova.compute.manager [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Terminating instance [ 1045.338489] env[62952]: DEBUG nova.compute.manager [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.338697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.339883] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02a4184-d6ad-4175-a859-cae057768435 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.343791] env[62952]: DEBUG oslo_vmware.api [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1045.343791] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52686c2e-85b3-cc39-a314-20595268407f" [ 1045.343791] env[62952]: _type = "Task" [ 1045.343791] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.349578] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.350166] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee033ca8-8e64-4161-9ea2-5a5a9737bdba {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.354980] env[62952]: DEBUG oslo_vmware.api [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52686c2e-85b3-cc39-a314-20595268407f, 'name': SearchDatastore_Task, 'duration_secs': 0.007293} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.355285] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.357311] env[62952]: DEBUG oslo_vmware.api [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1045.357311] env[62952]: value = "task-1367539" [ 1045.357311] env[62952]: _type = "Task" [ 1045.357311] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.365053] env[62952]: DEBUG oslo_vmware.api [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.525775] env[62952]: DEBUG oslo_concurrency.lockutils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.526153] env[62952]: DEBUG oslo_concurrency.lockutils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.526305] env[62952]: INFO nova.compute.manager [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Attaching volume ac6ea8d6-f93e-490a-9d4e-c3c6816c1351 to /dev/sdb [ 1045.556788] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367537, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.563024] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddc9fba-e5d4-4769-941c-ee0d030c9763 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.570575] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866b4774-2b44-40c4-b962-8f225e9f8d9d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.584994] env[62952]: INFO nova.compute.manager [-] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Took 1.24 seconds to deallocate network for instance. [ 1045.585596] env[62952]: DEBUG nova.virt.block_device [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updating existing volume attachment record: 55d0375c-2144-4782-88db-9830ec6887c5 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1045.588579] env[62952]: DEBUG nova.network.neutron [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [{"id": "aa4656a7-42d8-47ba-a29d-817000a4b596", "address": "fa:16:3e:14:80:1d", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa4656a7-42", "ovs_interfaceid": "aa4656a7-42d8-47ba-a29d-817000a4b596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.681529] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.789274] env[62952]: DEBUG nova.scheduler.client.report [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.874341] env[62952]: DEBUG oslo_vmware.api [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367539, 'name': PowerOffVM_Task, 'duration_secs': 0.494799} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.874749] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.875024] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.875500] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20c6cac9-08bc-41b7-af84-c1b01a49ea72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.926079] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "b9c57a92-73bc-4ee5-8431-2fe51a695383" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.926489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.060226] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367537, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.094030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "refresh_cache-88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.098377] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.179410] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367538, 'name': ReconfigVM_Task, 'duration_secs': 0.931321} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.179740] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfigured VM instance instance-0000005c to attach disk [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.180388] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4feda459-290f-4003-9766-8f92dee556ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.187594] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1046.187594] env[62952]: value = "task-1367544" [ 1046.187594] env[62952]: _type = "Task" [ 1046.187594] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.195893] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367544, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.296878] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.297522] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1046.300316] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.256s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.430714] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1046.559593] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367537, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.048149} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.559841] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] faa3b525-bd20-4e54-ba4c-60d65bf601aa/faa3b525-bd20-4e54-ba4c-60d65bf601aa.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.560100] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.560314] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1da5387-eb94-431f-8c39-56e8fe1cf4b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.568557] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1046.568557] env[62952]: value = "task-1367545" [ 1046.568557] env[62952]: _type = "Task" [ 1046.568557] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.576732] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.598034] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.598034] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5b5bc79-33a2-4a3b-8cb6-acdfc6ea0724 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.611968] env[62952]: DEBUG oslo_vmware.api [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1046.611968] env[62952]: value = "task-1367546" [ 1046.611968] env[62952]: _type = "Task" [ 1046.611968] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.621588] env[62952]: DEBUG oslo_vmware.api [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367546, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.697481] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367544, 'name': Rename_Task, 'duration_secs': 0.458454} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.697820] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.698092] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b16e30b-6385-4c8b-a559-db3321991999 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.705263] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1046.705263] env[62952]: value = "task-1367547" [ 1046.705263] env[62952]: _type = "Task" [ 1046.705263] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.713149] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367547, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.803922] env[62952]: DEBUG nova.compute.utils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.805597] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1046.805787] env[62952]: DEBUG nova.network.neutron [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.882271] env[62952]: DEBUG nova.policy [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1046.956367] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.080429] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078682} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.081475] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.081633] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941453ef-83d4-4902-a215-006991467061 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.107750] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] faa3b525-bd20-4e54-ba4c-60d65bf601aa/faa3b525-bd20-4e54-ba4c-60d65bf601aa.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.108490] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7376fe50-ab5b-4bf4-a765-5b3db7724d2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.137259] env[62952]: DEBUG oslo_vmware.api [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367546, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.138778] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1047.138778] env[62952]: value = "task-1367548" [ 1047.138778] env[62952]: _type = "Task" [ 1047.138778] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.149123] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367548, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.199923] env[62952]: DEBUG nova.network.neutron [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Successfully created port: a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1047.217939] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367547, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.314990] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1047.319507] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Applying migration context for instance e0384d0d-6738-47f7-b14d-994cae2d1fe4 as it has an incoming, in-progress migration d1159953-ebe6-4466-a646-a5de8ba8569f. Migration status is confirming {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1047.320841] env[62952]: INFO nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating resource usage from migration d1159953-ebe6-4466-a646-a5de8ba8569f [ 1047.343202] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.343435] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.343639] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.343885] env[62952]: WARNING nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ff189c9e-9e7c-4217-9c65-0f821393870a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1047.344108] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 532a2bcd-5d6a-4fa9-abc9-a048cc915fda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.344298] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 244f7b73-fda4-4477-b04d-c4f5cedd2d98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.344493] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 75d640ed-c41a-4761-8867-191d8b3e1f79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.344682] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Migration d1159953-ebe6-4466-a646-a5de8ba8569f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1047.344884] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance e0384d0d-6738-47f7-b14d-994cae2d1fe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.345075] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance f5e5c5c1-e2af-40e0-a957-67b0fd22718d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.345258] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance faa3b525-bd20-4e54-ba4c-60d65bf601aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.345467] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 12365f25-7d85-48dc-b2b7-9c8d0695e4a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1047.464205] env[62952]: DEBUG oslo_concurrency.lockutils [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.464502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.464700] env[62952]: DEBUG nova.compute.manager [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1047.465649] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec594b7f-2e3a-4a0a-bda5-738eb4ef0168 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.472978] env[62952]: DEBUG nova.compute.manager [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1047.473575] env[62952]: DEBUG nova.objects.instance [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'flavor' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.641219] env[62952]: DEBUG oslo_vmware.api [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367546, 'name': PowerOnVM_Task, 'duration_secs': 0.743555} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.646272] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1047.646440] env[62952]: DEBUG nova.compute.manager [None req-b2b3197a-cef1-4d52-9fb2-190b15478ec7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1047.647157] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3649684-f1f2-43b7-ad74-aeaa9efa5536 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.655528] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367548, 'name': ReconfigVM_Task, 'duration_secs': 0.284569} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.657323] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Reconfigured VM instance instance-0000005d to attach disk [datastore1] faa3b525-bd20-4e54-ba4c-60d65bf601aa/faa3b525-bd20-4e54-ba4c-60d65bf601aa.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.659946] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69623838-a3c6-4f10-91d7-9c1597b69da1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.669183] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1047.669183] env[62952]: value = "task-1367549" [ 1047.669183] env[62952]: _type = "Task" [ 1047.669183] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.677462] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367549, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.716834] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367547, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.850046] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance b9c57a92-73bc-4ee5-8431-2fe51a695383 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1047.850420] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1047.850885] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1047.978433] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.978720] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb4c5060-af83-4ef3-8b81-427f7da9b830 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.986133] env[62952]: DEBUG oslo_vmware.api [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1047.986133] env[62952]: value = "task-1367551" [ 1047.986133] env[62952]: _type = "Task" [ 1047.986133] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.995013] env[62952]: DEBUG oslo_vmware.api [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367551, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.016196] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f770aabb-8393-4218-b2ea-ed4e76537a73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.023290] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5257aa49-7b6e-4846-9e77-773f5d1df844 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.053121] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7ea511-6db7-4383-b5e4-a40e35b8d353 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.059867] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da80d8b0-0e0c-4bec-8fe3-c08643f395f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.072232] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.178604] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367549, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.217221] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367547, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.324564] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1048.349014] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.349307] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.349506] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.349700] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.349852] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.350007] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.350222] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.350392] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.350565] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.350805] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.351008] env[62952]: DEBUG nova.virt.hardware [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.351874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe90fd39-2ae0-42f0-99e5-1f483069d750 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.360552] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de9b80c-224b-42f7-8cdf-c147a692ffa5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.495443] env[62952]: DEBUG oslo_vmware.api [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367551, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.575650] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.608611] env[62952]: DEBUG nova.compute.manager [req-236391ef-045c-418f-b719-f2b8bdb53102 req-45e771cc-3299-4dea-b95b-94b842dd321c service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Received event network-vif-plugged-a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.608756] env[62952]: DEBUG oslo_concurrency.lockutils [req-236391ef-045c-418f-b719-f2b8bdb53102 req-45e771cc-3299-4dea-b95b-94b842dd321c service nova] Acquiring lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.608971] env[62952]: DEBUG oslo_concurrency.lockutils [req-236391ef-045c-418f-b719-f2b8bdb53102 req-45e771cc-3299-4dea-b95b-94b842dd321c service nova] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.609191] env[62952]: DEBUG oslo_concurrency.lockutils [req-236391ef-045c-418f-b719-f2b8bdb53102 req-45e771cc-3299-4dea-b95b-94b842dd321c service nova] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.609327] env[62952]: DEBUG nova.compute.manager [req-236391ef-045c-418f-b719-f2b8bdb53102 req-45e771cc-3299-4dea-b95b-94b842dd321c service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] No waiting events found dispatching network-vif-plugged-a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.609495] env[62952]: WARNING nova.compute.manager [req-236391ef-045c-418f-b719-f2b8bdb53102 req-45e771cc-3299-4dea-b95b-94b842dd321c service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Received unexpected event network-vif-plugged-a0a6224c-fa50-4b7d-a8cd-8cf619077179 for instance with vm_state building and task_state spawning. [ 1048.678786] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367549, 'name': Rename_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.720533] env[62952]: DEBUG oslo_vmware.api [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367547, 'name': PowerOnVM_Task, 'duration_secs': 1.711879} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.720533] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.720533] env[62952]: INFO nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Took 10.49 seconds to spawn the instance on the hypervisor. [ 1048.720533] env[62952]: DEBUG nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.720533] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad7e7a1-d030-4c1e-bebf-d08d08246bea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.770407] env[62952]: DEBUG nova.network.neutron [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Successfully updated port: a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.998366] env[62952]: DEBUG oslo_vmware.api [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367551, 'name': PowerOffVM_Task, 'duration_secs': 0.983179} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.998745] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.998908] env[62952]: DEBUG nova.compute.manager [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.999780] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f3d343-1a79-4f32-bfd3-45d34aa8a1a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.080394] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1049.080627] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.780s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.080900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.726s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.179489] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367549, 'name': Rename_Task, 'duration_secs': 1.168397} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.179796] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.180065] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f884b41-f56d-423a-a679-6be7c8eb91f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.188871] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1049.188871] env[62952]: value = "task-1367552" [ 1049.188871] env[62952]: _type = "Task" [ 1049.188871] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.198354] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.243210] env[62952]: INFO nova.compute.manager [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Took 15.53 seconds to build instance. [ 1049.273349] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-12365f25-7d85-48dc-b2b7-9c8d0695e4a9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.273746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-12365f25-7d85-48dc-b2b7-9c8d0695e4a9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.273876] env[62952]: DEBUG nova.network.neutron [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.299500] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.299778] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.299986] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleting the datastore file [datastore1] 244f7b73-fda4-4477-b04d-c4f5cedd2d98 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.300271] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-91f501be-f5f6-47bf-b61c-ae76e49ec0b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.309966] env[62952]: DEBUG oslo_vmware.api [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1049.309966] env[62952]: value = "task-1367553" [ 1049.309966] env[62952]: _type = "Task" [ 1049.309966] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.321038] env[62952]: DEBUG oslo_vmware.api [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367553, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.513876] env[62952]: DEBUG oslo_concurrency.lockutils [None req-05888dfb-d07f-4309-a16e-90840ba59e53 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.049s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.702090] env[62952]: DEBUG oslo_vmware.api [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367552, 'name': PowerOnVM_Task, 'duration_secs': 0.499996} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.702392] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1049.702596] env[62952]: INFO nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Took 9.12 seconds to spawn the instance on the hypervisor. [ 1049.703686] env[62952]: DEBUG nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.703686] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3d862d-8529-4ef9-bf07-ce6c12bd59ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.747307] env[62952]: DEBUG oslo_concurrency.lockutils [None req-142e5c51-9338-4ed2-8c34-32da8537f8b6 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.039s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.782495] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa688555-8b8f-40c9-a83a-082cf04bbb5e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.793745] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7fa4d1-07f7-429e-9142-28d6629b08a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.831394] env[62952]: DEBUG nova.network.neutron [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1049.837801] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b957fdcd-86a5-40e7-9ce5-830e02d328f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.848193] env[62952]: DEBUG oslo_vmware.api [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367553, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159114} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.850568] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.850878] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.851118] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.851313] env[62952]: INFO nova.compute.manager [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Took 4.51 seconds to destroy the instance on the hypervisor. [ 1049.851567] env[62952]: DEBUG oslo.service.loopingcall [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.851837] env[62952]: DEBUG nova.compute.manager [-] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1049.851968] env[62952]: DEBUG nova.network.neutron [-] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1049.855589] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b33b41-77cb-4a6d-90cd-d3dd817d88e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.876174] env[62952]: DEBUG nova.compute.provider_tree [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.107217] env[62952]: DEBUG nova.network.neutron [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Updating instance_info_cache with network_info: [{"id": "a0a6224c-fa50-4b7d-a8cd-8cf619077179", "address": "fa:16:3e:fe:96:c9", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0a6224c-fa", "ovs_interfaceid": "a0a6224c-fa50-4b7d-a8cd-8cf619077179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.138119] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1050.138414] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290968', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'name': 'volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '75d640ed-c41a-4761-8867-191d8b3e1f79', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'serial': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1050.139297] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30a244f-0c5a-4781-abaf-79e2da28fec8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.161407] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc970eae-40e7-46ad-83b1-280b56459006 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.188692] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351/volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.188988] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb64f84c-38a0-4a47-9329-c2b4e978dccd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.210368] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1050.210368] env[62952]: value = "task-1367554" [ 1050.210368] env[62952]: _type = "Task" [ 1050.210368] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.222928] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367554, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.225302] env[62952]: INFO nova.compute.manager [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Took 15.83 seconds to build instance. [ 1050.379696] env[62952]: DEBUG nova.scheduler.client.report [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.582247] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684ef6d4-86c3-4227-bc19-ccbfd5cb2bdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.589691] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Suspending the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1050.589942] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9c862686-c9d5-487a-ab0d-655d065b818f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.596937] env[62952]: DEBUG oslo_vmware.api [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1050.596937] env[62952]: value = "task-1367555" [ 1050.596937] env[62952]: _type = "Task" [ 1050.596937] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.606976] env[62952]: DEBUG oslo_vmware.api [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367555, 'name': SuspendVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.609690] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-12365f25-7d85-48dc-b2b7-9c8d0695e4a9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.610106] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Instance network_info: |[{"id": "a0a6224c-fa50-4b7d-a8cd-8cf619077179", "address": "fa:16:3e:fe:96:c9", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0a6224c-fa", "ovs_interfaceid": "a0a6224c-fa50-4b7d-a8cd-8cf619077179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1050.610591] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:96:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0a6224c-fa50-4b7d-a8cd-8cf619077179', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.619658] env[62952]: DEBUG oslo.service.loopingcall [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.619891] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1050.620139] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35702710-8812-414a-8839-6c9965570aef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.643042] env[62952]: DEBUG nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Received event network-changed-a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.643274] env[62952]: DEBUG nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Refreshing instance network info cache due to event network-changed-a0a6224c-fa50-4b7d-a8cd-8cf619077179. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1050.643509] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] Acquiring lock "refresh_cache-12365f25-7d85-48dc-b2b7-9c8d0695e4a9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.643669] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] Acquired lock "refresh_cache-12365f25-7d85-48dc-b2b7-9c8d0695e4a9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.643855] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Refreshing network info cache for port a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1050.647268] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.647268] env[62952]: value = "task-1367556" [ 1050.647268] env[62952]: _type = "Task" [ 1050.647268] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.660598] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367556, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.722273] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.727393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b3f9c51c-dd07-4de7-8fae-fb0db6a257ae tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.341s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.781132] env[62952]: DEBUG nova.network.neutron [-] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.832719] env[62952]: DEBUG nova.compute.manager [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Stashing vm_state: stopped {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1051.107362] env[62952]: DEBUG oslo_vmware.api [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367555, 'name': SuspendVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.158189] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367556, 'name': CreateVM_Task, 'duration_secs': 0.395675} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.158374] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1051.159095] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.159225] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.159552] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1051.159815] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-153dec8b-bf25-4956-a618-450611682409 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.165289] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1051.165289] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]522756c9-f2e2-e78b-b254-79e4b77f3e2a" [ 1051.165289] env[62952]: _type = "Task" [ 1051.165289] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.175141] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522756c9-f2e2-e78b-b254-79e4b77f3e2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.224287] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367554, 'name': ReconfigVM_Task, 'duration_secs': 0.859973} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.224475] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351/volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.231698] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdf72d71-cce6-4c65-b653-f9621ca8f038 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.251500] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1051.251500] env[62952]: value = "task-1367557" [ 1051.251500] env[62952]: _type = "Task" [ 1051.251500] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.262357] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367557, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.284064] env[62952]: INFO nova.compute.manager [-] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Took 1.43 seconds to deallocate network for instance. [ 1051.355783] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.389217] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.308s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.391944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.294s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.392157] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.394510] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.438s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.395964] env[62952]: INFO nova.compute.claims [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.417999] env[62952]: INFO nova.scheduler.client.report [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Deleted allocations for instance ff189c9e-9e7c-4217-9c65-0f821393870a [ 1051.449650] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Updated VIF entry in instance network info cache for port a0a6224c-fa50-4b7d-a8cd-8cf619077179. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1051.450028] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Updating instance_info_cache with network_info: [{"id": "a0a6224c-fa50-4b7d-a8cd-8cf619077179", "address": "fa:16:3e:fe:96:c9", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0a6224c-fa", "ovs_interfaceid": "a0a6224c-fa50-4b7d-a8cd-8cf619077179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.607429] env[62952]: DEBUG oslo_vmware.api [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367555, 'name': SuspendVM_Task, 'duration_secs': 0.634489} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.607643] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Suspended the VM {{(pid=62952) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1051.607825] env[62952]: DEBUG nova.compute.manager [None req-56dc8a38-e4ce-44bc-a701-4e953ce0a94d tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.608629] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f826b9-dc95-46c1-8e71-04394ea0f8d4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.677305] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522756c9-f2e2-e78b-b254-79e4b77f3e2a, 'name': SearchDatastore_Task, 'duration_secs': 0.011843} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.677594] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.677827] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.678076] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.678229] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.678412] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.678665] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c06b7cc6-de66-4999-ba9d-82201038e1c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.688603] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.688803] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.689887] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30c9fe3b-a757-49a6-9195-002abf2ed4dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.695802] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1051.695802] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5285d673-d8e0-9555-bc56-d68d7925a6e5" [ 1051.695802] env[62952]: _type = "Task" [ 1051.695802] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.703823] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5285d673-d8e0-9555-bc56-d68d7925a6e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.762253] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367557, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.791045] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.930237] env[62952]: DEBUG oslo_concurrency.lockutils [None req-af2d37f9-124a-44b4-b937-5e76b7a6fc6a tempest-AttachInterfacesTestJSON-1722434810 tempest-AttachInterfacesTestJSON-1722434810-project-member] Lock "ff189c9e-9e7c-4217-9c65-0f821393870a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.721s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.953062] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] Releasing lock "refresh_cache-12365f25-7d85-48dc-b2b7-9c8d0695e4a9" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.953729] env[62952]: DEBUG nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Received event network-changed-be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.953852] env[62952]: DEBUG nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Refreshing instance network info cache due to event network-changed-be83803d-9485-4459-882a-5dc9383e0f62. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.954222] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] Acquiring lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.954429] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] Acquired lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.954597] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Refreshing network info cache for port be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1051.957010] env[62952]: INFO nova.scheduler.client.report [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocation for migration d1159953-ebe6-4466-a646-a5de8ba8569f [ 1052.209370] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5285d673-d8e0-9555-bc56-d68d7925a6e5, 'name': SearchDatastore_Task, 'duration_secs': 0.008936} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.209370] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37f5a5e5-92ca-4061-be93-ea0d0aa99af3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.216181] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1052.216181] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ceccd7-7f95-9146-e857-f8e9c02b7906" [ 1052.216181] env[62952]: _type = "Task" [ 1052.216181] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.225038] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ceccd7-7f95-9146-e857-f8e9c02b7906, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.264054] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367557, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.463667] env[62952]: DEBUG oslo_concurrency.lockutils [None req-96860250-a332-4f0a-bf50-a75986fe2b4f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.949s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.579539] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e3c5bf-aa3f-4704-987a-899c0161e43d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.587770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5dca9a-115f-4a47-bb91-b042d74b0624 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.622858] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0399991-9d21-46dd-a2e5-f6a0ec299655 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.634339] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556352b7-36e6-4619-8af1-194384ff24c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.650350] env[62952]: DEBUG nova.compute.provider_tree [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.725991] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ceccd7-7f95-9146-e857-f8e9c02b7906, 'name': SearchDatastore_Task, 'duration_secs': 0.011217} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.726275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.726631] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 12365f25-7d85-48dc-b2b7-9c8d0695e4a9/12365f25-7d85-48dc-b2b7-9c8d0695e4a9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.726813] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-25031f86-d2ce-40cd-8799-2205f18c95fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.734916] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1052.734916] env[62952]: value = "task-1367558" [ 1052.734916] env[62952]: _type = "Task" [ 1052.734916] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.742774] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.764999] env[62952]: DEBUG oslo_vmware.api [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367557, 'name': ReconfigVM_Task, 'duration_secs': 1.157015} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.765338] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290968', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'name': 'volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '75d640ed-c41a-4761-8867-191d8b3e1f79', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'serial': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1052.793876] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updated VIF entry in instance network info cache for port be83803d-9485-4459-882a-5dc9383e0f62. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1052.794308] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating instance_info_cache with network_info: [{"id": "be83803d-9485-4459-882a-5dc9383e0f62", "address": "fa:16:3e:0b:bc:d7", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe83803d-94", "ovs_interfaceid": "be83803d-9485-4459-882a-5dc9383e0f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.130142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.130484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.131145] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.131444] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.131703] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.134119] env[62952]: INFO nova.compute.manager [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Terminating instance [ 1053.136568] env[62952]: DEBUG nova.compute.manager [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.136874] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1053.137762] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc90bc1-6515-4705-b36c-e8a0ba46dd31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.147795] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.148177] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9eade097-39d1-47e9-bfeb-d7c4414a3858 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.152105] env[62952]: DEBUG nova.scheduler.client.report [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.246694] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504556} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.247152] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 12365f25-7d85-48dc-b2b7-9c8d0695e4a9/12365f25-7d85-48dc-b2b7-9c8d0695e4a9.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1053.247260] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.247505] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ef774a0-22bb-4602-82ba-a2699a047035 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.251852] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.252062] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.252239] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleting the datastore file [datastore1] faa3b525-bd20-4e54-ba4c-60d65bf601aa {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.252490] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f503076-0b1d-4490-a370-bbf4471221ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.256950] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1053.256950] env[62952]: value = "task-1367560" [ 1053.256950] env[62952]: _type = "Task" [ 1053.256950] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.261213] env[62952]: DEBUG oslo_vmware.api [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1053.261213] env[62952]: value = "task-1367561" [ 1053.261213] env[62952]: _type = "Task" [ 1053.261213] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.273431] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367560, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.279872] env[62952]: DEBUG oslo_vmware.api [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.299441] env[62952]: DEBUG oslo_concurrency.lockutils [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] Releasing lock "refresh_cache-f5e5c5c1-e2af-40e0-a957-67b0fd22718d" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.299441] env[62952]: DEBUG nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Received event network-vif-deleted-fe087c06-3240-4db7-9b5d-ec61d2d983c0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.299441] env[62952]: INFO nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Neutron deleted interface fe087c06-3240-4db7-9b5d-ec61d2d983c0; detaching it from the instance and deleting it from the info cache [ 1053.299441] env[62952]: DEBUG nova.network.neutron [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.660018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.660018] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1053.660754] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.305s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.770557] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367560, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085249} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.770693] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.771704] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87ac581-de97-44d6-a719-4640a6a14088 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.778935] env[62952]: DEBUG oslo_vmware.api [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158654} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.780599] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.780599] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.780599] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.780599] env[62952]: INFO nova.compute.manager [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1053.780599] env[62952]: DEBUG oslo.service.loopingcall [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.780599] env[62952]: DEBUG nova.compute.manager [-] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.780943] env[62952]: DEBUG nova.network.neutron [-] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.806171] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 12365f25-7d85-48dc-b2b7-9c8d0695e4a9/12365f25-7d85-48dc-b2b7-9c8d0695e4a9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.807234] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b8c21c0-166a-45a1-a2d5-93e275f68ae5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.829357] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad27bb62-07dd-4caa-b549-a5c10796a39c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.836205] env[62952]: DEBUG nova.objects.instance [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'flavor' on Instance uuid 75d640ed-c41a-4761-8867-191d8b3e1f79 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.862697] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2988126f-c798-4c07-bafd-f800c2de0873 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.882208] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1053.882208] env[62952]: value = "task-1367562" [ 1053.882208] env[62952]: _type = "Task" [ 1053.882208] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.903119] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367562, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.923587] env[62952]: DEBUG nova.compute.manager [req-8a19eaee-ee0d-4ebd-b7d0-04d746e0eb17 req-9e1755e3-32a7-46c0-8a9f-5ea7f754c2f3 service nova] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Detach interface failed, port_id=fe087c06-3240-4db7-9b5d-ec61d2d983c0, reason: Instance 244f7b73-fda4-4477-b04d-c4f5cedd2d98 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.167222] env[62952]: DEBUG nova.compute.utils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.171775] env[62952]: INFO nova.compute.claims [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.177017] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.177017] env[62952]: DEBUG nova.network.neutron [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1054.180092] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.180369] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.180579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.180758] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.180927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.185485] env[62952]: INFO nova.compute.manager [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Terminating instance [ 1054.187448] env[62952]: DEBUG nova.compute.manager [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1054.187734] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.188567] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b6d6d8-2b96-4ef5-8d65-b11bf3a122dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.197858] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.198139] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53fba231-80fd-4ad2-9609-cefe2be7723c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.206248] env[62952]: DEBUG oslo_vmware.api [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1054.206248] env[62952]: value = "task-1367563" [ 1054.206248] env[62952]: _type = "Task" [ 1054.206248] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.216244] env[62952]: DEBUG oslo_vmware.api [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367563, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.289050] env[62952]: DEBUG nova.policy [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55ffb58a3b3a4ad29699d48ecd6c0700', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d564dc3c75e430dbaf1f90a9c90c18b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.335301] env[62952]: DEBUG nova.compute.manager [req-d5fbec89-9118-4c99-9120-9910011a1c3c req-1cafbc28-00fb-4e6e-8053-93d2e0e921ac service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Received event network-vif-deleted-9cf16459-ba22-47ad-82f5-735a585a9e9e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.335965] env[62952]: INFO nova.compute.manager [req-d5fbec89-9118-4c99-9120-9910011a1c3c req-1cafbc28-00fb-4e6e-8053-93d2e0e921ac service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Neutron deleted interface 9cf16459-ba22-47ad-82f5-735a585a9e9e; detaching it from the instance and deleting it from the info cache [ 1054.336272] env[62952]: DEBUG nova.network.neutron [req-d5fbec89-9118-4c99-9120-9910011a1c3c req-1cafbc28-00fb-4e6e-8053-93d2e0e921ac service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.345948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-431173a8-2675-4a34-84bc-3babf7273f6d tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.819s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.398894] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367562, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.586724] env[62952]: DEBUG nova.network.neutron [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Successfully created port: cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.678753] env[62952]: INFO nova.compute.resource_tracker [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating resource usage from migration dfe91621-c2c2-4e3e-90d4-a8a338e722ba [ 1054.682750] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.725968] env[62952]: DEBUG oslo_vmware.api [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367563, 'name': PowerOffVM_Task, 'duration_secs': 0.301981} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.726252] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1054.727101] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.727426] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5c52012-fb72-457d-bc8e-386a4b8f757b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.767850] env[62952]: DEBUG nova.network.neutron [-] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.804721] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.804972] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.808477] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore1] e0384d0d-6738-47f7-b14d-994cae2d1fe4 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.808477] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f48b418a-3d1e-4a2c-afa2-d03fb9e21643 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.815659] env[62952]: DEBUG oslo_vmware.api [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1054.815659] env[62952]: value = "task-1367565" [ 1054.815659] env[62952]: _type = "Task" [ 1054.815659] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.829236] env[62952]: DEBUG oslo_vmware.api [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367565, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.846690] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06aef431-2ad0-48b7-b83d-5fc3e41ddab9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.855194] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7d669b-a405-4327-a4a9-c9cdadb772ae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.893393] env[62952]: DEBUG nova.compute.manager [req-d5fbec89-9118-4c99-9120-9910011a1c3c req-1cafbc28-00fb-4e6e-8053-93d2e0e921ac service nova] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Detach interface failed, port_id=9cf16459-ba22-47ad-82f5-735a585a9e9e, reason: Instance faa3b525-bd20-4e54-ba4c-60d65bf601aa could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.906748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.907019] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.916218] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367562, 'name': ReconfigVM_Task, 'duration_secs': 0.640987} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.916616] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 12365f25-7d85-48dc-b2b7-9c8d0695e4a9/12365f25-7d85-48dc-b2b7-9c8d0695e4a9.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.920694] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98bce288-999b-4dbd-b116-ae21115803e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.928382] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1054.928382] env[62952]: value = "task-1367566" [ 1054.928382] env[62952]: _type = "Task" [ 1054.928382] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.940548] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367566, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.981100] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff0ad1c-9d1a-4978-98a4-92363d4fb13b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.989747] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14059eb-c206-4a60-a0c2-98652f147222 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.022928] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e9d283-6af1-43fd-9fcd-fb9683ac78fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.031933] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac02517f-b151-4aa1-bc11-143a1ef3f68a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.048274] env[62952]: DEBUG nova.compute.provider_tree [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.270439] env[62952]: INFO nova.compute.manager [-] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Took 1.49 seconds to deallocate network for instance. [ 1055.326466] env[62952]: DEBUG oslo_vmware.api [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367565, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270956} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.326947] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.327088] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.327274] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.327495] env[62952]: INFO nova.compute.manager [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1055.327776] env[62952]: DEBUG oslo.service.loopingcall [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.327995] env[62952]: DEBUG nova.compute.manager [-] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1055.328103] env[62952]: DEBUG nova.network.neutron [-] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1055.411514] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1055.439536] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367566, 'name': Rename_Task, 'duration_secs': 0.261999} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.440090] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1055.440428] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e08040e-1bc9-4716-ae98-34d2d77b41ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.448841] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1055.448841] env[62952]: value = "task-1367567" [ 1055.448841] env[62952]: _type = "Task" [ 1055.448841] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.458972] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.553325] env[62952]: DEBUG nova.scheduler.client.report [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.698308] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1055.720031] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.720031] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.720240] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.720425] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.720618] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.720810] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.721103] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.721287] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.721500] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.721711] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.721897] env[62952]: DEBUG nova.virt.hardware [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.722783] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c14258d-41d2-4343-afdc-89dd0e672232 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.732814] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ff7797-5a0b-4de8-843a-3f116a79d606 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.781971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.931177] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.961368] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367567, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.977932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "39eef6be-fe54-418b-b88d-104d082c3ca7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.978188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.058162] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.397s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.058380] env[62952]: INFO nova.compute.manager [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Migrating [ 1056.064927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.274s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.065214] env[62952]: DEBUG nova.objects.instance [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lazy-loading 'resources' on Instance uuid 244f7b73-fda4-4477-b04d-c4f5cedd2d98 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.161418] env[62952]: DEBUG nova.network.neutron [-] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.378285] env[62952]: DEBUG nova.compute.manager [req-66181d76-d6ab-4207-8ad0-1a3720d5b42f req-e046baa5-0a41-43ab-b29b-032f7e7e2364 service nova] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Received event network-vif-deleted-3f88a864-36dc-4fc8-8073-977444f51ed8 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.463705] env[62952]: DEBUG oslo_vmware.api [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367567, 'name': PowerOnVM_Task, 'duration_secs': 0.832791} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.464101] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1056.464476] env[62952]: INFO nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Took 8.14 seconds to spawn the instance on the hypervisor. [ 1056.464914] env[62952]: DEBUG nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.465820] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f029ff28-6dd3-4c67-bd0e-928781ddc20d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.480561] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.576828] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.576828] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.577240] env[62952]: DEBUG nova.network.neutron [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.651182] env[62952]: DEBUG nova.network.neutron [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Successfully updated port: cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.664773] env[62952]: INFO nova.compute.manager [-] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Took 1.34 seconds to deallocate network for instance. [ 1056.775560] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ca84a0-407b-4434-b2c6-4e71845e15ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.784487] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94c3f32-4a20-45f4-8a4d-6c17d9923057 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.816640] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747f526c-e85b-40c0-bbd8-6abb09b8a8a5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.824729] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d8cbbc-8835-41bb-9ac7-9fc8e208f7e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.838758] env[62952]: DEBUG nova.compute.provider_tree [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.983731] env[62952]: INFO nova.compute.manager [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Took 13.89 seconds to build instance. [ 1057.003550] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.153245] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "refresh_cache-b9c57a92-73bc-4ee5-8431-2fe51a695383" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.153389] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "refresh_cache-b9c57a92-73bc-4ee5-8431-2fe51a695383" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.153761] env[62952]: DEBUG nova.network.neutron [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.170748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.298463] env[62952]: DEBUG nova.network.neutron [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.341457] env[62952]: DEBUG nova.scheduler.client.report [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.486629] env[62952]: DEBUG oslo_concurrency.lockutils [None req-608fba82-a2a5-4574-8773-040a77b47ffc tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.405s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.683845] env[62952]: DEBUG nova.network.neutron [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1057.801484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.823955] env[62952]: DEBUG nova.network.neutron [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Updating instance_info_cache with network_info: [{"id": "cc951e8b-7cd0-4eef-86e6-7d760f8601ae", "address": "fa:16:3e:e6:6f:84", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc951e8b-7c", "ovs_interfaceid": "cc951e8b-7cd0-4eef-86e6-7d760f8601ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.847113] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.848193] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.066s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.848453] env[62952]: DEBUG nova.objects.instance [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lazy-loading 'resources' on Instance uuid faa3b525-bd20-4e54-ba4c-60d65bf601aa {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.866572] env[62952]: INFO nova.scheduler.client.report [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted allocations for instance 244f7b73-fda4-4477-b04d-c4f5cedd2d98 [ 1058.325395] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "refresh_cache-b9c57a92-73bc-4ee5-8431-2fe51a695383" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.325717] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Instance network_info: |[{"id": "cc951e8b-7cd0-4eef-86e6-7d760f8601ae", "address": "fa:16:3e:e6:6f:84", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc951e8b-7c", "ovs_interfaceid": "cc951e8b-7cd0-4eef-86e6-7d760f8601ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.326189] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:6f:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92fe29b3-0907-453d-aabb-5559c4bd7c0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc951e8b-7cd0-4eef-86e6-7d760f8601ae', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.334033] env[62952]: DEBUG oslo.service.loopingcall [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.334258] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.334482] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c6c7246-dd1e-46e5-9dbb-3b2d499fd2b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.356764] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.356764] env[62952]: value = "task-1367568" [ 1058.356764] env[62952]: _type = "Task" [ 1058.356764] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.364911] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367568, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.375397] env[62952]: DEBUG oslo_concurrency.lockutils [None req-458e4ef0-df65-4ea9-8151-4e4c37d5c3fc tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "244f7b73-fda4-4477-b04d-c4f5cedd2d98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.042s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.405399] env[62952]: DEBUG nova.compute.manager [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Received event network-vif-plugged-cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.405399] env[62952]: DEBUG oslo_concurrency.lockutils [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] Acquiring lock "b9c57a92-73bc-4ee5-8431-2fe51a695383-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.405609] env[62952]: DEBUG oslo_concurrency.lockutils [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.405725] env[62952]: DEBUG oslo_concurrency.lockutils [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.405896] env[62952]: DEBUG nova.compute.manager [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] No waiting events found dispatching network-vif-plugged-cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1058.406079] env[62952]: WARNING nova.compute.manager [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Received unexpected event network-vif-plugged-cc951e8b-7cd0-4eef-86e6-7d760f8601ae for instance with vm_state building and task_state spawning. [ 1058.406313] env[62952]: DEBUG nova.compute.manager [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Received event network-changed-cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.406486] env[62952]: DEBUG nova.compute.manager [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Refreshing instance network info cache due to event network-changed-cc951e8b-7cd0-4eef-86e6-7d760f8601ae. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1058.406669] env[62952]: DEBUG oslo_concurrency.lockutils [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] Acquiring lock "refresh_cache-b9c57a92-73bc-4ee5-8431-2fe51a695383" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.406806] env[62952]: DEBUG oslo_concurrency.lockutils [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] Acquired lock "refresh_cache-b9c57a92-73bc-4ee5-8431-2fe51a695383" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.406967] env[62952]: DEBUG nova.network.neutron [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Refreshing network info cache for port cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1058.528934] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100512ed-6f99-42be-b60a-4730083578af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.537961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6738b7-f0ed-439d-b92a-10a0922791fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.571824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6873dd6-7a42-44f1-8471-4fca4ff0a100 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.580063] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278bd5c8-56e6-4d76-bbd3-3687bae3aad1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.593276] env[62952]: DEBUG nova.compute.provider_tree [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.773959] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "44f0943d-7267-4472-8fe2-445a0727654a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.774222] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "44f0943d-7267-4472-8fe2-445a0727654a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.867880] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367568, 'name': CreateVM_Task, 'duration_secs': 0.32456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.868107] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.868768] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.868948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.869331] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.869607] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa419880-b287-45f4-a416-a676e3254501 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.875241] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1058.875241] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528538b1-935c-14d0-7e07-5e9dec722524" [ 1058.875241] env[62952]: _type = "Task" [ 1058.875241] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.886528] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528538b1-935c-14d0-7e07-5e9dec722524, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.096645] env[62952]: DEBUG nova.scheduler.client.report [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.152774] env[62952]: DEBUG nova.network.neutron [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Updated VIF entry in instance network info cache for port cc951e8b-7cd0-4eef-86e6-7d760f8601ae. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1059.153153] env[62952]: DEBUG nova.network.neutron [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Updating instance_info_cache with network_info: [{"id": "cc951e8b-7cd0-4eef-86e6-7d760f8601ae", "address": "fa:16:3e:e6:6f:84", "network": {"id": "4e7b94bf-e200-4adc-86f0-b633126e9957", "bridge": "br-int", "label": "tempest-ImagesTestJSON-930504367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d564dc3c75e430dbaf1f90a9c90c18b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc951e8b-7c", "ovs_interfaceid": "cc951e8b-7cd0-4eef-86e6-7d760f8601ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.277357] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1059.318059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70f6530-95d0-4bce-866e-e15570993fc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.336886] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1059.385804] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528538b1-935c-14d0-7e07-5e9dec722524, 'name': SearchDatastore_Task, 'duration_secs': 0.013915} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.386129] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.386380] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.386618] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.386770] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.386950] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.387213] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec15ec61-8b15-4cec-87fd-d0f317cc924c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.398454] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.398633] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1059.399328] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3edddef-2b9b-4c50-8fe9-37c6cb425a95 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.404428] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1059.404428] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524a2389-e228-4256-305b-83b6cb107ce2" [ 1059.404428] env[62952]: _type = "Task" [ 1059.404428] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.412046] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524a2389-e228-4256-305b-83b6cb107ce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.602529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.604899] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.674s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.606871] env[62952]: INFO nova.compute.claims [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.626052] env[62952]: INFO nova.scheduler.client.report [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocations for instance faa3b525-bd20-4e54-ba4c-60d65bf601aa [ 1059.655872] env[62952]: DEBUG oslo_concurrency.lockutils [req-6fcd9de6-8d60-4307-b305-d08bec72122b req-08846b59-47e4-4419-9ee1-3df186a9ca8c service nova] Releasing lock "refresh_cache-b9c57a92-73bc-4ee5-8431-2fe51a695383" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.795513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.842751] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.843100] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64e13077-0639-462b-999b-5f71051216dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.851947] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1059.851947] env[62952]: value = "task-1367569" [ 1059.851947] env[62952]: _type = "Task" [ 1059.851947] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.860124] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.914314] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524a2389-e228-4256-305b-83b6cb107ce2, 'name': SearchDatastore_Task, 'duration_secs': 0.013557} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.915066] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92036dfb-f2d7-41ce-badb-4fff7930cc91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.920124] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1059.920124] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52e1748c-df0e-0b51-7be1-6f8263ac5c7f" [ 1059.920124] env[62952]: _type = "Task" [ 1059.920124] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.928278] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52e1748c-df0e-0b51-7be1-6f8263ac5c7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.133440] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde5021b-c060-4fe1-827f-2725487c21ca tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "faa3b525-bd20-4e54-ba4c-60d65bf601aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.003s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.365066] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1060.365066] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1060.431188] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52e1748c-df0e-0b51-7be1-6f8263ac5c7f, 'name': SearchDatastore_Task, 'duration_secs': 0.011001} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.431877] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.431877] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] b9c57a92-73bc-4ee5-8431-2fe51a695383/b9c57a92-73bc-4ee5-8431-2fe51a695383.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1060.432096] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c1b717a-9c4e-4d22-9cf8-fd29e4b0635c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.439365] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1060.439365] env[62952]: value = "task-1367570" [ 1060.439365] env[62952]: _type = "Task" [ 1060.439365] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.448408] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367570, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.826535] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c62fbc1-a013-4312-b8c7-8eb458fefd3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.836627] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9ce2ed-139c-4fbd-8a41-1e5a46310064 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.868436] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde7fc90-91a1-41f0-add3-84150ae7de67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.873070] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.873262] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.873424] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.873609] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.873756] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.873907] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.874127] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.875020] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.875020] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.875020] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.875020] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.879897] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccc381e4-00c8-40cc-88c2-dbcf8d8d54b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.896977] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1c8c93-6d28-47f5-9ea4-3158dda5213f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.902340] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1060.902340] env[62952]: value = "task-1367571" [ 1060.902340] env[62952]: _type = "Task" [ 1060.902340] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.913671] env[62952]: DEBUG nova.compute.provider_tree [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.922671] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367571, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.951034] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367570, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481648} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.951335] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] b9c57a92-73bc-4ee5-8431-2fe51a695383/b9c57a92-73bc-4ee5-8431-2fe51a695383.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.951561] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.951822] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dfbf6452-6691-40fc-ad5f-800b67977816 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.958567] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1060.958567] env[62952]: value = "task-1367572" [ 1060.958567] env[62952]: _type = "Task" [ 1060.958567] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.967487] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.367988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.367988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.412631] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367571, 'name': ReconfigVM_Task, 'duration_secs': 0.149244} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.412886] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1061.416681] env[62952]: DEBUG nova.scheduler.client.report [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.469261] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069534} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.469539] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.470364] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d43b3a-6f19-4905-99dd-15ddb5f2ceb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.492448] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] b9c57a92-73bc-4ee5-8431-2fe51a695383/b9c57a92-73bc-4ee5-8431-2fe51a695383.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.492940] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36596386-5d23-45b7-8398-84d67707d0a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.512046] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1061.512046] env[62952]: value = "task-1367573" [ 1061.512046] env[62952]: _type = "Task" [ 1061.512046] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.519925] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.869616] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1061.921891] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.922713] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.928913] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.929186] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.929319] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.929506] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.929659] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.929807] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.930028] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.930277] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.930461] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.930630] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.930804] env[62952]: DEBUG nova.virt.hardware [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.936432] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Reconfiguring VM instance instance-0000004a to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1061.936957] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.934s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.938572] env[62952]: INFO nova.compute.claims [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1061.941156] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b51a4af4-91d3-443a-a986-bd1bcd81b82c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.962830] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1061.962830] env[62952]: value = "task-1367574" [ 1061.962830] env[62952]: _type = "Task" [ 1061.962830] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.974564] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.027532] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.400124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.439293] env[62952]: DEBUG nova.compute.utils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.440711] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1062.440882] env[62952]: DEBUG nova.network.neutron [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1062.473691] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367574, 'name': ReconfigVM_Task, 'duration_secs': 0.179721} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.473989] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Reconfigured VM instance instance-0000004a to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1062.474814] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ab36b7-e9da-491f-8a25-11aadf9ac56f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.497430] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2/867ffed2-50ec-45d8-b64e-989cf8d1b0e2.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.499054] env[62952]: DEBUG nova.policy [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17b50c91ccd347e4b45126a8f23c0278', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9657a113032417fb4b97716a772f8ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1062.500469] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18ab2db4-5dc3-4c23-9b13-340f0898e139 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.522940] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367573, 'name': ReconfigVM_Task, 'duration_secs': 0.832703} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.524457] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Reconfigured VM instance instance-0000005f to attach disk [datastore1] b9c57a92-73bc-4ee5-8431-2fe51a695383/b9c57a92-73bc-4ee5-8431-2fe51a695383.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.525212] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1062.525212] env[62952]: value = "task-1367575" [ 1062.525212] env[62952]: _type = "Task" [ 1062.525212] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.525461] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c11a33c-b8b1-4bf9-8928-53dcb23bbfeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.536110] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367575, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.537405] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1062.537405] env[62952]: value = "task-1367576" [ 1062.537405] env[62952]: _type = "Task" [ 1062.537405] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.546204] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367576, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.809201] env[62952]: DEBUG nova.network.neutron [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Successfully created port: 6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.944627] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1063.038049] env[62952]: DEBUG oslo_vmware.api [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367575, 'name': ReconfigVM_Task, 'duration_secs': 0.281002} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.044594] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2/867ffed2-50ec-45d8-b64e-989cf8d1b0e2.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.044942] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1063.055032] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367576, 'name': Rename_Task, 'duration_secs': 0.155451} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.055335] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.055592] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d93238f-5110-4b82-be42-16624d12a7ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.064673] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1063.064673] env[62952]: value = "task-1367577" [ 1063.064673] env[62952]: _type = "Task" [ 1063.064673] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.073496] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.176197] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edbe3dd-63d8-4dfd-b381-ce1aa1f2d2fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.183952] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623cf055-8dd4-4b4f-9c7f-f9e0b900512a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.216985] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdf3c1e-f51d-481a-b60e-d4517aecc787 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.225194] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40b6d0e-5d23-4036-9c3b-cef84d19ee6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.239373] env[62952]: DEBUG nova.compute.provider_tree [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.552092] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85254b2b-9748-4e49-b70b-9c8f19cda349 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.575475] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64931467-3e50-4891-bd88-9926fceba109 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.583446] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367577, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.596625] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1063.743070] env[62952]: DEBUG nova.scheduler.client.report [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.958935] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.986388] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.986679] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.986852] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.987085] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.987212] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.987355] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.987564] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.987760] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.988357] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.988357] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.988357] env[62952]: DEBUG nova.virt.hardware [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.989229] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982a920c-0145-48e4-88c2-d0acecc0133b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.997616] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a5a132-0e75-4d7f-91ed-a72097427187 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.080976] env[62952]: DEBUG oslo_vmware.api [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367577, 'name': PowerOnVM_Task, 'duration_secs': 0.875013} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.081266] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.081477] env[62952]: INFO nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Took 8.38 seconds to spawn the instance on the hypervisor. [ 1064.081661] env[62952]: DEBUG nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.082462] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc86d3b-bc70-44db-bb5b-1300fab9ed07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.154894] env[62952]: DEBUG nova.network.neutron [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Port 4b40f3e8-7283-4edc-8016-d2a0bde62f60 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1064.248290] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.251023] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1064.252167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.081s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.252525] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.254730] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.459s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.256514] env[62952]: INFO nova.compute.claims [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1064.282991] env[62952]: INFO nova.scheduler.client.report [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocations for instance e0384d0d-6738-47f7-b14d-994cae2d1fe4 [ 1064.339420] env[62952]: DEBUG nova.compute.manager [req-1df77564-2f6f-4e54-8198-a74b2eedb03b req-6a167a16-7f18-4a40-ada5-6d2d736b8c8a service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Received event network-vif-plugged-6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.339621] env[62952]: DEBUG oslo_concurrency.lockutils [req-1df77564-2f6f-4e54-8198-a74b2eedb03b req-6a167a16-7f18-4a40-ada5-6d2d736b8c8a service nova] Acquiring lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.339893] env[62952]: DEBUG oslo_concurrency.lockutils [req-1df77564-2f6f-4e54-8198-a74b2eedb03b req-6a167a16-7f18-4a40-ada5-6d2d736b8c8a service nova] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.340104] env[62952]: DEBUG oslo_concurrency.lockutils [req-1df77564-2f6f-4e54-8198-a74b2eedb03b req-6a167a16-7f18-4a40-ada5-6d2d736b8c8a service nova] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.340287] env[62952]: DEBUG nova.compute.manager [req-1df77564-2f6f-4e54-8198-a74b2eedb03b req-6a167a16-7f18-4a40-ada5-6d2d736b8c8a service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] No waiting events found dispatching network-vif-plugged-6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1064.340459] env[62952]: WARNING nova.compute.manager [req-1df77564-2f6f-4e54-8198-a74b2eedb03b req-6a167a16-7f18-4a40-ada5-6d2d736b8c8a service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Received unexpected event network-vif-plugged-6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 for instance with vm_state building and task_state spawning. [ 1064.427255] env[62952]: DEBUG nova.network.neutron [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Successfully updated port: 6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.601856] env[62952]: INFO nova.compute.manager [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Took 17.67 seconds to build instance. [ 1064.761446] env[62952]: DEBUG nova.compute.utils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1064.764714] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1064.764828] env[62952]: DEBUG nova.network.neutron [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1064.791250] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55295e25-e945-4393-9ad9-382dc7cd23b0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "e0384d0d-6738-47f7-b14d-994cae2d1fe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.611s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.808135] env[62952]: DEBUG nova.policy [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e67193d02201461bb7e4339b5025ca48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfc390d64c0463190f071f3f62936dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1064.931461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.931461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.931648] env[62952]: DEBUG nova.network.neutron [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.102489] env[62952]: DEBUG nova.network.neutron [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Successfully created port: a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1065.104653] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9f036c48-1ee3-4188-a176-e307e864f25a tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.178s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.178731] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.178989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.179194] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.265104] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1065.467413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85747fea-0dcd-4b30-8b25-a4251ca1e745 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.470608] env[62952]: DEBUG nova.network.neutron [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1065.477363] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef74f23c-0b30-4e8b-a282-578121e19358 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.510391] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdadf656-15f4-424b-ba16-9f26724ff8ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.519025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770a7aba-35de-41eb-adb7-e7e87bca6623 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.534149] env[62952]: DEBUG nova.compute.provider_tree [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.631138] env[62952]: DEBUG nova.network.neutron [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [{"id": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "address": "fa:16:3e:74:a7:d4", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bbebc32-dc", "ovs_interfaceid": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.939262] env[62952]: DEBUG nova.compute.manager [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.940237] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979aee9a-d8ff-453a-a069-94ad7409bb44 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.039249] env[62952]: DEBUG nova.scheduler.client.report [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.133782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.134129] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Instance network_info: |[{"id": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "address": "fa:16:3e:74:a7:d4", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bbebc32-dc", "ovs_interfaceid": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1066.134558] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:a7:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.141982] env[62952]: DEBUG oslo.service.loopingcall [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.142214] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.142435] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2027455e-7de7-42ac-a7e6-8eb989a98003 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.162181] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.162181] env[62952]: value = "task-1367578" [ 1066.162181] env[62952]: _type = "Task" [ 1066.162181] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.169835] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367578, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.214464] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.214648] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.214827] env[62952]: DEBUG nova.network.neutron [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.277359] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1066.301813] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1066.302127] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1066.302294] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.302481] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1066.302634] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.302782] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1066.303079] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1066.303266] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1066.303471] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1066.303658] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1066.303835] env[62952]: DEBUG nova.virt.hardware [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.304781] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc1f7c8-037c-4f0c-baea-46510b243331 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.312528] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8100bf3-1c0b-4269-8c46-5cf6cbf5eef9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.366145] env[62952]: DEBUG nova.compute.manager [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Received event network-changed-6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.366554] env[62952]: DEBUG nova.compute.manager [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Refreshing instance network info cache due to event network-changed-6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.366925] env[62952]: DEBUG oslo_concurrency.lockutils [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] Acquiring lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.367269] env[62952]: DEBUG oslo_concurrency.lockutils [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] Acquired lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.367725] env[62952]: DEBUG nova.network.neutron [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Refreshing network info cache for port 6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.452424] env[62952]: INFO nova.compute.manager [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] instance snapshotting [ 1066.455981] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f033445-4e0c-42a3-bfe1-a09e0cb2f68e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.477932] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb56f53-7cbb-4f50-be88-71154974dc24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.543789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.544733] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1066.547008] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.147s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.548425] env[62952]: INFO nova.compute.claims [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.672298] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367578, 'name': CreateVM_Task, 'duration_secs': 0.312228} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.672469] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.673167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.673339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.673677] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.673934] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c693a88-c3d7-4a4b-9347-8dd2a0880318 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.678766] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1066.678766] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]522d829b-c27d-b87d-1175-916356282a7c" [ 1066.678766] env[62952]: _type = "Task" [ 1066.678766] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.686187] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522d829b-c27d-b87d-1175-916356282a7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.725805] env[62952]: DEBUG nova.network.neutron [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Successfully updated port: a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1066.989135] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1066.989762] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-09cac18e-a1fb-40f9-88be-581f422fd1a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.997976] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1066.997976] env[62952]: value = "task-1367579" [ 1066.997976] env[62952]: _type = "Task" [ 1066.997976] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.002119] env[62952]: DEBUG nova.network.neutron [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.009927] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367579, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.060301] env[62952]: DEBUG nova.compute.utils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.065650] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1067.065650] env[62952]: DEBUG nova.network.neutron [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1067.126329] env[62952]: DEBUG nova.policy [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1067.158029] env[62952]: DEBUG nova.network.neutron [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updated VIF entry in instance network info cache for port 6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.158432] env[62952]: DEBUG nova.network.neutron [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [{"id": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "address": "fa:16:3e:74:a7:d4", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bbebc32-dc", "ovs_interfaceid": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.189610] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522d829b-c27d-b87d-1175-916356282a7c, 'name': SearchDatastore_Task, 'duration_secs': 0.010174} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.189920] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.190166] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.190915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.190915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.190915] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.191103] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7edbad13-eed0-44ab-aaba-4b127bca4f14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.199190] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.199390] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.200106] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-441e3979-ffa8-4f5e-8404-1c3f712af502 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.205037] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1067.205037] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52510066-e1d5-232d-4c27-93381f302b45" [ 1067.205037] env[62952]: _type = "Task" [ 1067.205037] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.214399] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52510066-e1d5-232d-4c27-93381f302b45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.228070] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.228207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.228470] env[62952]: DEBUG nova.network.neutron [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1067.392788] env[62952]: DEBUG nova.network.neutron [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Successfully created port: df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1067.507867] env[62952]: DEBUG oslo_concurrency.lockutils [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.511410] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367579, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.568104] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1067.661210] env[62952]: DEBUG oslo_concurrency.lockutils [req-afaff5d9-7450-40aa-ae3a-87ada1cdb753 req-32ce594e-702d-475a-918f-784477ffb9f6 service nova] Releasing lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.720016] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52510066-e1d5-232d-4c27-93381f302b45, 'name': SearchDatastore_Task, 'duration_secs': 0.010017} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.720861] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5ccb284-8542-4287-805d-21580312a3e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.726257] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1067.726257] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52361577-620d-8858-50d7-ad60a736505f" [ 1067.726257] env[62952]: _type = "Task" [ 1067.726257] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.740125] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52361577-620d-8858-50d7-ad60a736505f, 'name': SearchDatastore_Task, 'duration_secs': 0.009493} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.740241] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.740671] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2/ea621ff2-e54e-4403-9e6d-dea84c7e8ad2.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1067.740939] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-274064ea-315a-421e-8486-ad44dc09b388 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.748751] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1067.748751] env[62952]: value = "task-1367580" [ 1067.748751] env[62952]: _type = "Task" [ 1067.748751] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.759624] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.764633] env[62952]: DEBUG nova.network.neutron [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1067.768108] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8a59c2-9ba5-42e0-ba3f-7451b888fcfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.774790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384a4777-dfaa-46b7-8822-8e44d7b5a99c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.808310] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a5bfd4-fbaa-4ed2-8bdd-e092283ae762 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.816340] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c031d3e8-8785-4491-8ceb-0a811bd4fcc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.833024] env[62952]: DEBUG nova.compute.provider_tree [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.010332] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367579, 'name': CreateSnapshot_Task, 'duration_secs': 0.837313} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.010682] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1068.011502] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10506ba-1bec-434c-8196-0890e64aa18b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.021303] env[62952]: DEBUG nova.network.neutron [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Updating instance_info_cache with network_info: [{"id": "a6a41344-0103-472c-a33e-a85dc9674a09", "address": "fa:16:3e:88:db:ff", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a41344-01", "ovs_interfaceid": "a6a41344-0103-472c-a33e-a85dc9674a09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.032792] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbe29b5-2182-46aa-8287-6fba0204acad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.058966] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfd538d-c997-4db1-beb0-de210aab6ebb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.068098] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.259113] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461461} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.259483] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2/ea621ff2-e54e-4403-9e6d-dea84c7e8ad2.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1068.259619] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.259861] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ec708a7-80b2-4bee-8a0e-f972e76cc0c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.266996] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1068.266996] env[62952]: value = "task-1367581" [ 1068.266996] env[62952]: _type = "Task" [ 1068.266996] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.275998] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.336372] env[62952]: DEBUG nova.scheduler.client.report [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.392389] env[62952]: DEBUG nova.compute.manager [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Received event network-vif-plugged-a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.392490] env[62952]: DEBUG oslo_concurrency.lockutils [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] Acquiring lock "39eef6be-fe54-418b-b88d-104d082c3ca7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.392784] env[62952]: DEBUG oslo_concurrency.lockutils [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.393039] env[62952]: DEBUG oslo_concurrency.lockutils [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.393231] env[62952]: DEBUG nova.compute.manager [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] No waiting events found dispatching network-vif-plugged-a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1068.393494] env[62952]: WARNING nova.compute.manager [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Received unexpected event network-vif-plugged-a6a41344-0103-472c-a33e-a85dc9674a09 for instance with vm_state building and task_state spawning. [ 1068.393666] env[62952]: DEBUG nova.compute.manager [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Received event network-changed-a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.393825] env[62952]: DEBUG nova.compute.manager [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Refreshing instance network info cache due to event network-changed-a6a41344-0103-472c-a33e-a85dc9674a09. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1068.393998] env[62952]: DEBUG oslo_concurrency.lockutils [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] Acquiring lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.529758] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.530113] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Instance network_info: |[{"id": "a6a41344-0103-472c-a33e-a85dc9674a09", "address": "fa:16:3e:88:db:ff", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a41344-01", "ovs_interfaceid": "a6a41344-0103-472c-a33e-a85dc9674a09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1068.537012] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1068.537323] env[62952]: DEBUG oslo_concurrency.lockutils [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] Acquired lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.537508] env[62952]: DEBUG nova.network.neutron [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Refreshing network info cache for port a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1068.538791] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:db:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6a41344-0103-472c-a33e-a85dc9674a09', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1068.546389] env[62952]: DEBUG oslo.service.loopingcall [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.546565] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-06726716-52b8-40a5-8cea-0a45c749c046 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.551846] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1068.552312] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4393b7b-09a2-4459-8905-8aa93b5c2db2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.575025] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6082d1-44a6-4061-bd6a-306f5c8239fb tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance '867ffed2-50ec-45d8-b64e-989cf8d1b0e2' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.578114] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1068.578114] env[62952]: value = "task-1367582" [ 1068.578114] env[62952]: _type = "Task" [ 1068.578114] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.580018] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1068.581833] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1068.581833] env[62952]: value = "task-1367583" [ 1068.581833] env[62952]: _type = "Task" [ 1068.581833] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.593379] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367583, 'name': CreateVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.596112] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367582, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.607123] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.607382] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.607597] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.607728] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.607879] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.608038] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.608275] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.608447] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.608664] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.608790] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.608964] env[62952]: DEBUG nova.virt.hardware [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.610026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc684ca-9f62-4214-ada4-b3e3b817b54f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.620480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c4e100-c091-445c-a722-258d21062b4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.779947] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066588} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.782252] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.783088] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdd466e-c3cc-4bd2-8822-0f908a882ae5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.805147] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2/ea621ff2-e54e-4403-9e6d-dea84c7e8ad2.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.805474] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4aef7c80-22b2-47b4-95d3-79b64ac62b5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.825492] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1068.825492] env[62952]: value = "task-1367584" [ 1068.825492] env[62952]: _type = "Task" [ 1068.825492] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.833491] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367584, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.842381] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.842899] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1068.904096] env[62952]: DEBUG nova.network.neutron [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Updated VIF entry in instance network info cache for port a6a41344-0103-472c-a33e-a85dc9674a09. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.904477] env[62952]: DEBUG nova.network.neutron [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Updating instance_info_cache with network_info: [{"id": "a6a41344-0103-472c-a33e-a85dc9674a09", "address": "fa:16:3e:88:db:ff", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a41344-01", "ovs_interfaceid": "a6a41344-0103-472c-a33e-a85dc9674a09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.039282] env[62952]: DEBUG nova.network.neutron [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Successfully updated port: df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1069.098378] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367582, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.103201] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367583, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.336554] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367584, 'name': ReconfigVM_Task, 'duration_secs': 0.317524} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.336912] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Reconfigured VM instance instance-00000060 to attach disk [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2/ea621ff2-e54e-4403-9e6d-dea84c7e8ad2.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.337633] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2003760-7f02-43af-97e5-db395a191d76 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.344533] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1069.344533] env[62952]: value = "task-1367585" [ 1069.344533] env[62952]: _type = "Task" [ 1069.344533] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.348305] env[62952]: DEBUG nova.compute.utils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1069.349590] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1069.349759] env[62952]: DEBUG nova.network.neutron [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1069.356479] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367585, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.390961] env[62952]: DEBUG nova.policy [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f24a30e1160a43d5be6cde39b8f79edd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '094ca56818604bee9c1f75d1493ab3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1069.407596] env[62952]: DEBUG oslo_concurrency.lockutils [req-50124a8f-0906-4aa1-8691-cdc8e5da2abc req-e632eef4-7fc8-4330-bb20-50fe7d701c5b service nova] Releasing lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.545498] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-44f0943d-7267-4472-8fe2-445a0727654a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.545498] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-44f0943d-7267-4472-8fe2-445a0727654a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.545498] env[62952]: DEBUG nova.network.neutron [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.602068] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367583, 'name': CreateVM_Task, 'duration_secs': 0.77424} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.602513] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367582, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.602743] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1069.603501] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.603831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.604093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1069.604387] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd88c150-f395-4356-b4d8-12d6e7a496af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.609094] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1069.609094] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5241fda2-9ecd-0e36-72b3-de5fbef031d8" [ 1069.609094] env[62952]: _type = "Task" [ 1069.609094] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.618377] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5241fda2-9ecd-0e36-72b3-de5fbef031d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.670895] env[62952]: DEBUG nova.network.neutron [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Successfully created port: 69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.855053] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1069.857883] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367585, 'name': Rename_Task, 'duration_secs': 0.137698} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.857883] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1069.858463] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76761ff3-4c51-4fc8-84df-ba94e10ce340 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.867724] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1069.867724] env[62952]: value = "task-1367586" [ 1069.867724] env[62952]: _type = "Task" [ 1069.867724] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.876490] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367586, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.076051] env[62952]: DEBUG nova.network.neutron [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1070.102999] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367582, 'name': CloneVM_Task, 'duration_secs': 1.353748} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.104171] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Created linked-clone VM from snapshot [ 1070.104995] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61be95ee-e1bf-46a5-94b8-856a52263175 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.112631] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Uploading image 54d5d42f-bda9-4fe2-a950-f9ed7a9815ea {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1070.126738] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5241fda2-9ecd-0e36-72b3-de5fbef031d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009882} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.126836] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1070.127182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.127410] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.127689] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.127846] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.128039] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.128275] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-66a5477b-7c86-4af5-88d9-bc88e1c3fe50 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.130018] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-973fb3d8-e993-4736-a76e-8554df6c82ac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.137451] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1070.137451] env[62952]: value = "task-1367587" [ 1070.137451] env[62952]: _type = "Task" [ 1070.137451] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.138498] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.138677] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1070.142239] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d2e35ec-c371-4b9e-b7f6-8fbf332ce207 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.154305] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367587, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.154621] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1070.154621] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c8fabb-1d39-a2ab-4dad-bcd04a093afb" [ 1070.154621] env[62952]: _type = "Task" [ 1070.154621] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.163150] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c8fabb-1d39-a2ab-4dad-bcd04a093afb, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.163872] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a47ffad-a885-416f-9a5f-735e79288f17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.168865] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1070.168865] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524d049d-b9c1-0ba5-d321-acacfcd62d1e" [ 1070.168865] env[62952]: _type = "Task" [ 1070.168865] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.177846] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524d049d-b9c1-0ba5-d321-acacfcd62d1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.245986] env[62952]: DEBUG nova.network.neutron [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Updating instance_info_cache with network_info: [{"id": "df27cdc8-1bd7-46ad-9cfe-38ab3802d917", "address": "fa:16:3e:b6:e9:b5", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf27cdc8-1b", "ovs_interfaceid": "df27cdc8-1bd7-46ad-9cfe-38ab3802d917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.382022] env[62952]: DEBUG oslo_vmware.api [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367586, 'name': PowerOnVM_Task, 'duration_secs': 0.469789} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.382022] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1070.382022] env[62952]: INFO nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Took 6.42 seconds to spawn the instance on the hypervisor. [ 1070.382022] env[62952]: DEBUG nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.382022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba58727-e0a8-42be-bae3-1bd5bbb7077a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.471680] env[62952]: DEBUG nova.compute.manager [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Received event network-vif-plugged-df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.471999] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] Acquiring lock "44f0943d-7267-4472-8fe2-445a0727654a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.472573] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] Lock "44f0943d-7267-4472-8fe2-445a0727654a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.472930] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] Lock "44f0943d-7267-4472-8fe2-445a0727654a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.473301] env[62952]: DEBUG nova.compute.manager [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] No waiting events found dispatching network-vif-plugged-df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.473504] env[62952]: WARNING nova.compute.manager [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Received unexpected event network-vif-plugged-df27cdc8-1bd7-46ad-9cfe-38ab3802d917 for instance with vm_state building and task_state spawning. [ 1070.473880] env[62952]: DEBUG nova.compute.manager [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Received event network-changed-df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.474123] env[62952]: DEBUG nova.compute.manager [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Refreshing instance network info cache due to event network-changed-df27cdc8-1bd7-46ad-9cfe-38ab3802d917. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1070.474330] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] Acquiring lock "refresh_cache-44f0943d-7267-4472-8fe2-445a0727654a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.647916] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367587, 'name': Destroy_Task} progress is 33%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.678790] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]524d049d-b9c1-0ba5-d321-acacfcd62d1e, 'name': SearchDatastore_Task, 'duration_secs': 0.008345} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.679090] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.679392] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 39eef6be-fe54-418b-b88d-104d082c3ca7/39eef6be-fe54-418b-b88d-104d082c3ca7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1070.679741] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b8ee662-7f5f-4860-81c0-1cabc9b622f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.687567] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1070.687567] env[62952]: value = "task-1367588" [ 1070.687567] env[62952]: _type = "Task" [ 1070.687567] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.695443] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.748961] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-44f0943d-7267-4472-8fe2-445a0727654a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.749251] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Instance network_info: |[{"id": "df27cdc8-1bd7-46ad-9cfe-38ab3802d917", "address": "fa:16:3e:b6:e9:b5", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf27cdc8-1b", "ovs_interfaceid": "df27cdc8-1bd7-46ad-9cfe-38ab3802d917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1070.749562] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] Acquired lock "refresh_cache-44f0943d-7267-4472-8fe2-445a0727654a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.749763] env[62952]: DEBUG nova.network.neutron [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Refreshing network info cache for port df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1070.751449] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:e9:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df27cdc8-1bd7-46ad-9cfe-38ab3802d917', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.759853] env[62952]: DEBUG oslo.service.loopingcall [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.762617] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.763128] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6dc19a4-bd8f-4e1e-8bf1-3f1f8f9198a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.786682] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.786682] env[62952]: value = "task-1367589" [ 1070.786682] env[62952]: _type = "Task" [ 1070.786682] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.795569] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367589, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.839221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.839628] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.839899] env[62952]: DEBUG nova.compute.manager [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Going to confirm migration 2 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1070.869293] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1070.900874] env[62952]: INFO nova.compute.manager [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Took 14.98 seconds to build instance. [ 1070.903893] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.904177] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.904297] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.904486] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.904637] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.904790] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.905009] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.905319] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.905828] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.905828] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.906071] env[62952]: DEBUG nova.virt.hardware [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.906901] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e5e47a-7d54-4727-b003-8501f23d804c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.925353] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa1dc65-1419-4564-8ae3-7e059053484f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.156017] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367587, 'name': Destroy_Task, 'duration_secs': 0.643663} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.158817] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Destroyed the VM [ 1071.159140] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1071.159482] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-11760cc8-d3a4-4a35-a02f-502b2c26a750 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.170850] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1071.170850] env[62952]: value = "task-1367590" [ 1071.170850] env[62952]: _type = "Task" [ 1071.170850] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.180967] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367590, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.199877] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367588, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.305580] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367589, 'name': CreateVM_Task, 'duration_secs': 0.46524} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.305838] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.306656] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.306831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.307161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.307493] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-338a21c7-1408-4f97-969b-026c6da623fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.313653] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1071.313653] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528d8d62-b397-73e8-2b1d-58f246250278" [ 1071.313653] env[62952]: _type = "Task" [ 1071.313653] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.323560] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528d8d62-b397-73e8-2b1d-58f246250278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.407321] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.407634] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.407888] env[62952]: DEBUG nova.network.neutron [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.408411] env[62952]: DEBUG nova.objects.instance [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'info_cache' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.413031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-745fdfb3-2af9-4130-8d99-48de0d49a378 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.506s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.434634] env[62952]: DEBUG nova.network.neutron [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Updated VIF entry in instance network info cache for port df27cdc8-1bd7-46ad-9cfe-38ab3802d917. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1071.435128] env[62952]: DEBUG nova.network.neutron [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Updating instance_info_cache with network_info: [{"id": "df27cdc8-1bd7-46ad-9cfe-38ab3802d917", "address": "fa:16:3e:b6:e9:b5", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf27cdc8-1b", "ovs_interfaceid": "df27cdc8-1bd7-46ad-9cfe-38ab3802d917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.599535] env[62952]: DEBUG nova.network.neutron [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Successfully updated port: 69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1071.682584] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367590, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.699238] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542598} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.699647] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 39eef6be-fe54-418b-b88d-104d082c3ca7/39eef6be-fe54-418b-b88d-104d082c3ca7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1071.699746] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.699955] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b43c0ec6-6378-4288-9236-4d3185a180f7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.706340] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1071.706340] env[62952]: value = "task-1367591" [ 1071.706340] env[62952]: _type = "Task" [ 1071.706340] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.714247] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367591, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.823607] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528d8d62-b397-73e8-2b1d-58f246250278, 'name': SearchDatastore_Task, 'duration_secs': 0.009195} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.823913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.824174] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.824424] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.824576] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.824756] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.825051] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5889600-d7c5-43bd-aa34-c8cc9618fc39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.833013] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.833013] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.833627] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-139043c7-3244-4116-a3c8-018c05038633 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.838384] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1071.838384] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ce02c1-5614-667e-104b-62220c6238c7" [ 1071.838384] env[62952]: _type = "Task" [ 1071.838384] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.847113] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ce02c1-5614-667e-104b-62220c6238c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.939560] env[62952]: DEBUG oslo_concurrency.lockutils [req-71ae5449-71c6-4ca7-b0ad-9e882e9e481f req-86f40e95-1e5a-462f-8219-253f6b147b5a service nova] Releasing lock "refresh_cache-44f0943d-7267-4472-8fe2-445a0727654a" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.101083] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.101263] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.101478] env[62952]: DEBUG nova.network.neutron [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1072.181927] env[62952]: DEBUG oslo_vmware.api [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367590, 'name': RemoveSnapshot_Task, 'duration_secs': 0.746648} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.182231] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1072.221040] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367591, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072035} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.221040] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.221722] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87ad6a4-4ae4-40ff-b8d4-28155a8e4786 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.244354] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 39eef6be-fe54-418b-b88d-104d082c3ca7/39eef6be-fe54-418b-b88d-104d082c3ca7.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.245512] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58dd5b6e-fcec-4533-8de5-2c88494d08e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.265359] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1072.265359] env[62952]: value = "task-1367592" [ 1072.265359] env[62952]: _type = "Task" [ 1072.265359] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.274982] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367592, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.348475] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ce02c1-5614-667e-104b-62220c6238c7, 'name': SearchDatastore_Task, 'duration_secs': 0.008867} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.349296] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eb36f0e-b637-402f-8837-f80525f4bad8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.354508] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1072.354508] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b5b16e-4e2f-14c9-44b5-10830f89192a" [ 1072.354508] env[62952]: _type = "Task" [ 1072.354508] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.362171] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b5b16e-4e2f-14c9-44b5-10830f89192a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.497895] env[62952]: DEBUG nova.compute.manager [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Received event network-vif-plugged-69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.498225] env[62952]: DEBUG oslo_concurrency.lockutils [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] Acquiring lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.498665] env[62952]: DEBUG oslo_concurrency.lockutils [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.498855] env[62952]: DEBUG oslo_concurrency.lockutils [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.499089] env[62952]: DEBUG nova.compute.manager [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] No waiting events found dispatching network-vif-plugged-69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1072.499297] env[62952]: WARNING nova.compute.manager [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Received unexpected event network-vif-plugged-69251416-59fd-45c5-a937-fe98301eb5a7 for instance with vm_state building and task_state spawning. [ 1072.499469] env[62952]: DEBUG nova.compute.manager [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Received event network-changed-69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.499622] env[62952]: DEBUG nova.compute.manager [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Refreshing instance network info cache due to event network-changed-69251416-59fd-45c5-a937-fe98301eb5a7. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.499793] env[62952]: DEBUG oslo_concurrency.lockutils [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] Acquiring lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.643450] env[62952]: DEBUG nova.network.neutron [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1072.662287] env[62952]: DEBUG nova.network.neutron [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.686961] env[62952]: WARNING nova.compute.manager [None req-569ea197-1d0a-4d0f-95cd-572913701e98 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Image not found during snapshot: nova.exception.ImageNotFound: Image 54d5d42f-bda9-4fe2-a950-f9ed7a9815ea could not be found. [ 1072.723308] env[62952]: DEBUG nova.compute.manager [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1072.779037] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367592, 'name': ReconfigVM_Task, 'duration_secs': 0.289055} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.781511] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 39eef6be-fe54-418b-b88d-104d082c3ca7/39eef6be-fe54-418b-b88d-104d082c3ca7.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.782754] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f47c4efe-9d1f-47c7-8a7b-aac4833b7872 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.789472] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1072.789472] env[62952]: value = "task-1367593" [ 1072.789472] env[62952]: _type = "Task" [ 1072.789472] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.799819] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367593, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.824247] env[62952]: DEBUG nova.network.neutron [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance_info_cache with network_info: [{"id": "69251416-59fd-45c5-a937-fe98301eb5a7", "address": "fa:16:3e:f6:6b:d5", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69251416-59", "ovs_interfaceid": "69251416-59fd-45c5-a937-fe98301eb5a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.865775] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b5b16e-4e2f-14c9-44b5-10830f89192a, 'name': SearchDatastore_Task, 'duration_secs': 0.013012} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.866034] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.866326] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 44f0943d-7267-4472-8fe2-445a0727654a/44f0943d-7267-4472-8fe2-445a0727654a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1072.866796] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58deef2d-c125-4ef6-9264-ae862d2d52d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.873097] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1072.873097] env[62952]: value = "task-1367594" [ 1072.873097] env[62952]: _type = "Task" [ 1072.873097] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.880906] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.165944] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.166267] env[62952]: DEBUG nova.objects.instance [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'migration_context' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.247735] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.248052] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.298949] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367593, 'name': Rename_Task, 'duration_secs': 0.132789} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.299246] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.299478] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a673ae3b-dc08-4916-96c3-45a3a6c2ddd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.306241] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1073.306241] env[62952]: value = "task-1367595" [ 1073.306241] env[62952]: _type = "Task" [ 1073.306241] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.314148] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.326971] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.327651] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Instance network_info: |[{"id": "69251416-59fd-45c5-a937-fe98301eb5a7", "address": "fa:16:3e:f6:6b:d5", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69251416-59", "ovs_interfaceid": "69251416-59fd-45c5-a937-fe98301eb5a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1073.328069] env[62952]: DEBUG oslo_concurrency.lockutils [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] Acquired lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.328597] env[62952]: DEBUG nova.network.neutron [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Refreshing network info cache for port 69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1073.330030] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:6b:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97113f46-d648-4613-b233-069acba18198', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69251416-59fd-45c5-a937-fe98301eb5a7', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1073.338469] env[62952]: DEBUG oslo.service.loopingcall [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.341658] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1073.342246] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80a19b66-8ab5-472e-a4e3-eeb838b33d66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.363025] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1073.363025] env[62952]: value = "task-1367596" [ 1073.363025] env[62952]: _type = "Task" [ 1073.363025] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.371615] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367596, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.382183] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367594, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.587403] env[62952]: DEBUG nova.network.neutron [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updated VIF entry in instance network info cache for port 69251416-59fd-45c5-a937-fe98301eb5a7. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.587791] env[62952]: DEBUG nova.network.neutron [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance_info_cache with network_info: [{"id": "69251416-59fd-45c5-a937-fe98301eb5a7", "address": "fa:16:3e:f6:6b:d5", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69251416-59", "ovs_interfaceid": "69251416-59fd-45c5-a937-fe98301eb5a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.669335] env[62952]: DEBUG nova.objects.base [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Object Instance<867ffed2-50ec-45d8-b64e-989cf8d1b0e2> lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1073.670271] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80cb607-2b4c-45b4-8ab7-318beaf1b4a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.692396] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c181c488-b45e-4755-9d56-352b0af278cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.697823] env[62952]: DEBUG oslo_vmware.api [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1073.697823] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ced9d2-e41a-a583-c3d2-e3a2e642dc7d" [ 1073.697823] env[62952]: _type = "Task" [ 1073.697823] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.705559] env[62952]: DEBUG oslo_vmware.api [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ced9d2-e41a-a583-c3d2-e3a2e642dc7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.753451] env[62952]: INFO nova.compute.claims [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.816922] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367595, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.873080] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367596, 'name': CreateVM_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.881970] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528136} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.882285] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 44f0943d-7267-4472-8fe2-445a0727654a/44f0943d-7267-4472-8fe2-445a0727654a.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.882499] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.882752] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-deeedb2e-74fa-45ad-af73-377e70e454d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.890256] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1073.890256] env[62952]: value = "task-1367597" [ 1073.890256] env[62952]: _type = "Task" [ 1073.890256] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.897811] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367597, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.934222] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "b9c57a92-73bc-4ee5-8431-2fe51a695383" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.934582] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.934909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "b9c57a92-73bc-4ee5-8431-2fe51a695383-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.935167] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.935407] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.937762] env[62952]: INFO nova.compute.manager [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Terminating instance [ 1073.939694] env[62952]: DEBUG nova.compute.manager [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.939952] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1073.940934] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fabf94-6847-4dab-af49-5bfbccff144e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.949077] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.949310] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82e21988-1465-4e3d-87c3-d9c945999125 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.955457] env[62952]: DEBUG oslo_vmware.api [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1073.955457] env[62952]: value = "task-1367598" [ 1073.955457] env[62952]: _type = "Task" [ 1073.955457] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.962925] env[62952]: DEBUG oslo_vmware.api [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.090598] env[62952]: DEBUG oslo_concurrency.lockutils [req-73bee72a-c1cc-4bd2-a539-c50cc4aee772 req-40f39b39-aeb5-48d3-951a-f18dd8683bd3 service nova] Releasing lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.208143] env[62952]: DEBUG oslo_vmware.api [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ced9d2-e41a-a583-c3d2-e3a2e642dc7d, 'name': SearchDatastore_Task, 'duration_secs': 0.00651} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.208455] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.259564] env[62952]: INFO nova.compute.resource_tracker [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating resource usage from migration 3effe577-bf29-41de-a612-7938583e828a [ 1074.317590] env[62952]: DEBUG oslo_vmware.api [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367595, 'name': PowerOnVM_Task, 'duration_secs': 0.880178} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.319987] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.320175] env[62952]: INFO nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Took 8.04 seconds to spawn the instance on the hypervisor. [ 1074.320364] env[62952]: DEBUG nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.321563] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb9051b-77ba-4b9d-9521-9a0ca84ce877 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.373642] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367596, 'name': CreateVM_Task, 'duration_secs': 0.723061} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.375777] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1074.376660] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.376766] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.377093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1074.377346] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80b9f603-c696-4d63-afff-85f5c9fcd324 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.382069] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1074.382069] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528c43f1-4e31-219c-c5af-9b6a1b45ec9f" [ 1074.382069] env[62952]: _type = "Task" [ 1074.382069] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.393636] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528c43f1-4e31-219c-c5af-9b6a1b45ec9f, 'name': SearchDatastore_Task, 'duration_secs': 0.009587} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.396556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.396789] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1074.397034] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.397188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.397366] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1074.399543] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2477755f-0c7d-45c7-9719-cb940f1bc50f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.406015] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367597, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073585} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.406346] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.407872] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e3e30e-8f88-40ce-83dc-bfe491da9469 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.410037] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1074.410214] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1074.410861] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74e75a50-98a0-4462-9ca5-3f342c6a161a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.419460] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1074.419460] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a1da54-d223-d235-3035-f8fb0c62bbed" [ 1074.419460] env[62952]: _type = "Task" [ 1074.419460] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.436882] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 44f0943d-7267-4472-8fe2-445a0727654a/44f0943d-7267-4472-8fe2-445a0727654a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.441660] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44bc9be6-283c-4432-b1fa-78699b42a35f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.469324] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a1da54-d223-d235-3035-f8fb0c62bbed, 'name': SearchDatastore_Task, 'duration_secs': 0.009038} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.472346] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434d9501-64f6-4d8a-9b63-b2d75fb41b91 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.474589] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1074.474589] env[62952]: value = "task-1367599" [ 1074.474589] env[62952]: _type = "Task" [ 1074.474589] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.477291] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e7c1b8c-69d4-4166-8414-050f32c652e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.479298] env[62952]: DEBUG oslo_vmware.api [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367598, 'name': PowerOffVM_Task, 'duration_secs': 0.214221} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.482411] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.482589] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.484502] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0be2ba01-b346-496b-83a1-4808c74c4a36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.486705] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e35955-e85d-4d45-88fa-3b961da4e209 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.493518] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367599, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.493833] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1074.493833] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52f07528-5702-e7a8-146b-09a07a57f17d" [ 1074.493833] env[62952]: _type = "Task" [ 1074.493833] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.523924] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008f4b1d-e996-4626-b689-0173138a49e9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.526715] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52f07528-5702-e7a8-146b-09a07a57f17d, 'name': SearchDatastore_Task, 'duration_secs': 0.011054} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.526975] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.527252] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed/9c18c8c7-bc97-4a84-a86d-e23f8cce20ed.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.527814] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8deea235-d673-4f93-add8-847cdabdf564 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.532675] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b8093b-3a41-47bc-af31-512409eae486 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.537016] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1074.537016] env[62952]: value = "task-1367601" [ 1074.537016] env[62952]: _type = "Task" [ 1074.537016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.549099] env[62952]: DEBUG nova.compute.provider_tree [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.556668] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.561031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.561031] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.561202] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleting the datastore file [datastore1] b9c57a92-73bc-4ee5-8431-2fe51a695383 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.561661] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9178c30f-2b1f-45dd-98bb-804baa8d3b6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.567585] env[62952]: DEBUG oslo_vmware.api [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for the task: (returnval){ [ 1074.567585] env[62952]: value = "task-1367602" [ 1074.567585] env[62952]: _type = "Task" [ 1074.567585] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.575120] env[62952]: DEBUG oslo_vmware.api [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.839067] env[62952]: INFO nova.compute.manager [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Took 17.85 seconds to build instance. [ 1074.989603] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367599, 'name': ReconfigVM_Task, 'duration_secs': 0.457225} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.989921] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 44f0943d-7267-4472-8fe2-445a0727654a/44f0943d-7267-4472-8fe2-445a0727654a.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.990609] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1951c1fc-2249-4c89-879a-b7eb48ea4e18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.997764] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1074.997764] env[62952]: value = "task-1367603" [ 1074.997764] env[62952]: _type = "Task" [ 1074.997764] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.011811] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367603, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.047440] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367601, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.052555] env[62952]: DEBUG nova.scheduler.client.report [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.078141] env[62952]: DEBUG oslo_vmware.api [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Task: {'id': task-1367602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215309} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.078389] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.078572] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.078755] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.078940] env[62952]: INFO nova.compute.manager [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1075.079219] env[62952]: DEBUG oslo.service.loopingcall [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.079403] env[62952]: DEBUG nova.compute.manager [-] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1075.079496] env[62952]: DEBUG nova.network.neutron [-] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1075.339778] env[62952]: DEBUG nova.compute.manager [req-a5759826-e808-439a-99c6-4abb8a1c8187 req-75b92f94-2e03-411e-9c45-06e6198faeab service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Received event network-vif-deleted-cc951e8b-7cd0-4eef-86e6-7d760f8601ae {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.339933] env[62952]: INFO nova.compute.manager [req-a5759826-e808-439a-99c6-4abb8a1c8187 req-75b92f94-2e03-411e-9c45-06e6198faeab service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Neutron deleted interface cc951e8b-7cd0-4eef-86e6-7d760f8601ae; detaching it from the instance and deleting it from the info cache [ 1075.340189] env[62952]: DEBUG nova.network.neutron [req-a5759826-e808-439a-99c6-4abb8a1c8187 req-75b92f94-2e03-411e-9c45-06e6198faeab service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.341689] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ab561cb3-a1aa-49f4-b23d-3a52b54f33ca tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.363s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.508539] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367603, 'name': Rename_Task, 'duration_secs': 0.34707} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.508539] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.508746] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5407fff-c35d-4aa7-b7d1-2a0c500870fd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.515392] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1075.515392] env[62952]: value = "task-1367604" [ 1075.515392] env[62952]: _type = "Task" [ 1075.515392] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.522579] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367604, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.547185] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543861} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.547436] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed/9c18c8c7-bc97-4a84-a86d-e23f8cce20ed.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1075.547657] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.547907] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07736efe-291c-4c7b-bafd-d9e4b9f05680 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.555031] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1075.555031] env[62952]: value = "task-1367605" [ 1075.555031] env[62952]: _type = "Task" [ 1075.555031] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.557835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.310s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.558062] env[62952]: INFO nova.compute.manager [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Migrating [ 1075.564787] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.356s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.575897] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.830106] env[62952]: DEBUG nova.network.neutron [-] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.843153] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a4312f6-a756-408b-8535-582de085824d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.853834] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25593cd-e45d-466c-99ec-288b5d484500 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.882721] env[62952]: DEBUG nova.compute.manager [req-a5759826-e808-439a-99c6-4abb8a1c8187 req-75b92f94-2e03-411e-9c45-06e6198faeab service nova] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Detach interface failed, port_id=cc951e8b-7cd0-4eef-86e6-7d760f8601ae, reason: Instance b9c57a92-73bc-4ee5-8431-2fe51a695383 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1076.026124] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367604, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.065450] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.077547] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.077768] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.077959] env[62952]: DEBUG nova.network.neutron [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1076.267032] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3a2659-5359-4245-b6ed-5e5e55685cbf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.275140] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c36d7b9-c3b1-4b87-8de4-596bb784d714 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.306028] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd1ddb5-c5ea-43c9-b991-7e073e21854b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.313690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f198512-e4a6-4b02-a3cd-9dc551e5a717 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.328556] env[62952]: INFO nova.compute.manager [-] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Took 1.25 seconds to deallocate network for instance. [ 1076.329024] env[62952]: DEBUG nova.compute.provider_tree [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.527564] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367604, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.565452] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.855944} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.565810] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1076.567650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865fc51b-1c34-46f2-946e-9be215c646a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.589333] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed/9c18c8c7-bc97-4a84-a86d-e23f8cce20ed.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.591591] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e76a2d98-1dd9-415a-ba50-e192c1e44aa0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.612031] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1076.612031] env[62952]: value = "task-1367606" [ 1076.612031] env[62952]: _type = "Task" [ 1076.612031] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.619992] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367606, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.835040] env[62952]: DEBUG nova.scheduler.client.report [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.838806] env[62952]: DEBUG nova.network.neutron [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [{"id": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "address": "fa:16:3e:74:a7:d4", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bbebc32-dc", "ovs_interfaceid": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.839561] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.026645] env[62952]: DEBUG oslo_vmware.api [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367604, 'name': PowerOnVM_Task, 'duration_secs': 1.284599} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.028064] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1077.028064] env[62952]: INFO nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1077.028064] env[62952]: DEBUG nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1077.028512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d282de86-dce6-4b67-84a2-26ec09ef98d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.122011] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367606, 'name': ReconfigVM_Task, 'duration_secs': 0.290011} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.122322] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed/9c18c8c7-bc97-4a84-a86d-e23f8cce20ed.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.122969] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e19e9227-bb37-4918-a6b4-280aec5e2124 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.129270] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1077.129270] env[62952]: value = "task-1367607" [ 1077.129270] env[62952]: _type = "Task" [ 1077.129270] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.137786] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367607, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.344719] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.366450] env[62952]: DEBUG nova.compute.manager [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Received event network-changed-a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.366658] env[62952]: DEBUG nova.compute.manager [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Refreshing instance network info cache due to event network-changed-a6a41344-0103-472c-a33e-a85dc9674a09. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1077.368765] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] Acquiring lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.368931] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] Acquired lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.369129] env[62952]: DEBUG nova.network.neutron [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Refreshing network info cache for port a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.544957] env[62952]: INFO nova.compute.manager [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Took 17.76 seconds to build instance. [ 1077.639565] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367607, 'name': Rename_Task, 'duration_secs': 0.265674} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.639844] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.640103] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-760b22ca-a83b-4725-a4a8-9ceec388e8b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.646215] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1077.646215] env[62952]: value = "task-1367608" [ 1077.646215] env[62952]: _type = "Task" [ 1077.646215] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.653577] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.848507] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.284s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.848784] env[62952]: DEBUG nova.compute.manager [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62952) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1077.851680] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.012s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.851904] env[62952]: DEBUG nova.objects.instance [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lazy-loading 'resources' on Instance uuid b9c57a92-73bc-4ee5-8431-2fe51a695383 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.047180] env[62952]: DEBUG oslo_concurrency.lockutils [None req-75eb9fca-ffd3-41e2-ae67-a67f8ca29ca8 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "44f0943d-7267-4472-8fe2-445a0727654a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.273s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.122051] env[62952]: DEBUG nova.network.neutron [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Updated VIF entry in instance network info cache for port a6a41344-0103-472c-a33e-a85dc9674a09. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1078.122607] env[62952]: DEBUG nova.network.neutron [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Updating instance_info_cache with network_info: [{"id": "a6a41344-0103-472c-a33e-a85dc9674a09", "address": "fa:16:3e:88:db:ff", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a41344-01", "ovs_interfaceid": "a6a41344-0103-472c-a33e-a85dc9674a09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.156307] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367608, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.431529] env[62952]: INFO nova.scheduler.client.report [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted allocation for migration dfe91621-c2c2-4e3e-90d4-a8a338e722ba [ 1078.553988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5db4cc-d415-4963-ad9e-0a7e1bd29fed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.564264] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0bfff6-8f46-4bb8-a537-f30d0f78c785 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.596026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375e37a5-8b73-4153-8eb6-a6938487470b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.602941] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f130b392-36f8-4f86-b687-019284b39419 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.616531] env[62952]: DEBUG nova.compute.provider_tree [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.625320] env[62952]: DEBUG oslo_concurrency.lockutils [req-cb90be34-fe3d-4d95-91b1-6f48a1e82d0e req-4332c1ef-ae0e-4b1c-9bfb-36e22cfc80ab service nova] Releasing lock "refresh_cache-39eef6be-fe54-418b-b88d-104d082c3ca7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.654999] env[62952]: DEBUG oslo_vmware.api [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367608, 'name': PowerOnVM_Task, 'duration_secs': 0.794303} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.655282] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.655529] env[62952]: INFO nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Took 7.79 seconds to spawn the instance on the hypervisor. [ 1078.655723] env[62952]: DEBUG nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.656719] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88008e5-b44e-45f9-9e17-edfd8a37b886 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.796132] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "44f0943d-7267-4472-8fe2-445a0727654a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.796425] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "44f0943d-7267-4472-8fe2-445a0727654a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.796695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "44f0943d-7267-4472-8fe2-445a0727654a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.796827] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "44f0943d-7267-4472-8fe2-445a0727654a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.796996] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "44f0943d-7267-4472-8fe2-445a0727654a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.799180] env[62952]: INFO nova.compute.manager [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Terminating instance [ 1078.801165] env[62952]: DEBUG nova.compute.manager [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.801452] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1078.802431] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2743164a-c643-47de-bb3c-201f3b4eabaf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.810808] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.811062] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5323e566-30b1-45f2-b921-28d2d0fa4da3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.817539] env[62952]: DEBUG oslo_vmware.api [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1078.817539] env[62952]: value = "task-1367609" [ 1078.817539] env[62952]: _type = "Task" [ 1078.817539] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.825522] env[62952]: DEBUG oslo_vmware.api [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367609, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.870754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86563e2-5897-4aae-8184-5cee64d98a49 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.890749] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1078.938602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-dde6efa3-c22f-4b4b-ad2a-2283cfbd6389 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.099s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.064940] env[62952]: DEBUG nova.objects.instance [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'flavor' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.119614] env[62952]: DEBUG nova.scheduler.client.report [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.173802] env[62952]: INFO nova.compute.manager [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Took 16.80 seconds to build instance. [ 1079.327195] env[62952]: DEBUG oslo_vmware.api [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367609, 'name': PowerOffVM_Task, 'duration_secs': 0.23536} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.327495] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.327655] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1079.327916] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e77ada7a-8238-40d5-af54-637b8e5875be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.396594] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.396904] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcf8ccda-7937-4a58-8302-09b8d73b5eeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.400396] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1079.400725] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1079.400993] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore1] 44f0943d-7267-4472-8fe2-445a0727654a {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.401677] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-274a944e-1c81-44e0-a3e4-c71e00fa52e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.406496] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1079.406496] env[62952]: value = "task-1367611" [ 1079.406496] env[62952]: _type = "Task" [ 1079.406496] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.410519] env[62952]: DEBUG oslo_vmware.api [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1079.410519] env[62952]: value = "task-1367612" [ 1079.410519] env[62952]: _type = "Task" [ 1079.410519] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.416391] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.421242] env[62952]: DEBUG oslo_vmware.api [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367612, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.444762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.444762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.444762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.444953] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.445142] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.447229] env[62952]: INFO nova.compute.manager [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Terminating instance [ 1079.449100] env[62952]: DEBUG nova.compute.manager [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1079.449284] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1079.450123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee503107-5180-4739-b814-e2a967afcffe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.457849] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.458109] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78f7dcb3-f632-48f5-ab52-e4edc6c25d0f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.464694] env[62952]: DEBUG oslo_vmware.api [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1079.464694] env[62952]: value = "task-1367613" [ 1079.464694] env[62952]: _type = "Task" [ 1079.464694] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.472475] env[62952]: DEBUG oslo_vmware.api [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367613, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.570330] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.570695] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.571039] env[62952]: DEBUG nova.network.neutron [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.571428] env[62952]: DEBUG nova.objects.instance [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'info_cache' on Instance uuid 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.624529] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.652450] env[62952]: INFO nova.scheduler.client.report [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Deleted allocations for instance b9c57a92-73bc-4ee5-8431-2fe51a695383 [ 1079.675853] env[62952]: DEBUG oslo_concurrency.lockutils [None req-011f0dd4-f4d9-4bfd-9eb5-bddd25a530a2 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.308s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.920747] env[62952]: DEBUG oslo_vmware.api [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367612, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163654} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.924071] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.924263] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.924466] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.924650] env[62952]: INFO nova.compute.manager [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1079.924913] env[62952]: DEBUG oslo.service.loopingcall [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.925152] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367611, 'name': PowerOffVM_Task, 'duration_secs': 0.189503} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.925358] env[62952]: DEBUG nova.compute.manager [-] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.925453] env[62952]: DEBUG nova.network.neutron [-] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1079.927041] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.927316] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.974908] env[62952]: DEBUG oslo_vmware.api [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367613, 'name': PowerOffVM_Task, 'duration_secs': 0.202447} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.975190] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.975363] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1079.975619] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5492f6d3-d8cf-4ab0-ae32-05353773adbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.047833] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.047954] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.048154] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Deleting the datastore file [datastore2] 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.048426] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be3ed358-2b26-4937-a457-4d8e485360a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.055514] env[62952]: DEBUG oslo_vmware.api [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1080.055514] env[62952]: value = "task-1367615" [ 1080.055514] env[62952]: _type = "Task" [ 1080.055514] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.062839] env[62952]: DEBUG oslo_vmware.api [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.075601] env[62952]: DEBUG nova.objects.base [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Object Instance<867ffed2-50ec-45d8-b64e-989cf8d1b0e2> lazy-loaded attributes: flavor,info_cache {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1080.166831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-da398fef-a5de-410a-9b3a-6e6ecf59f461 tempest-ImagesTestJSON-2007206826 tempest-ImagesTestJSON-2007206826-project-member] Lock "b9c57a92-73bc-4ee5-8431-2fe51a695383" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.229s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.201627] env[62952]: DEBUG nova.compute.manager [req-db1c79ff-1e34-493e-9044-1eb81c1141fd req-3cb72ae8-7537-4390-af77-cb8237d65eb2 service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Received event network-vif-deleted-df27cdc8-1bd7-46ad-9cfe-38ab3802d917 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.201900] env[62952]: INFO nova.compute.manager [req-db1c79ff-1e34-493e-9044-1eb81c1141fd req-3cb72ae8-7537-4390-af77-cb8237d65eb2 service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Neutron deleted interface df27cdc8-1bd7-46ad-9cfe-38ab3802d917; detaching it from the instance and deleting it from the info cache [ 1080.202105] env[62952]: DEBUG nova.network.neutron [req-db1c79ff-1e34-493e-9044-1eb81c1141fd req-3cb72ae8-7537-4390-af77-cb8237d65eb2 service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.434441] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.434441] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.434651] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.434651] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.434812] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.434942] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.435182] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.435487] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.435593] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.435721] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.435899] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.441439] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58cabbc7-0980-4543-8542-ac64df29bbde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.458780] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1080.458780] env[62952]: value = "task-1367616" [ 1080.458780] env[62952]: _type = "Task" [ 1080.458780] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.468241] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367616, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.566320] env[62952]: DEBUG oslo_vmware.api [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367615, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151992} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.566656] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1080.566843] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1080.567043] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1080.567264] env[62952]: INFO nova.compute.manager [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1080.567602] env[62952]: DEBUG oslo.service.loopingcall [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.567908] env[62952]: DEBUG nova.compute.manager [-] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1080.568016] env[62952]: DEBUG nova.network.neutron [-] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1080.663486] env[62952]: DEBUG nova.network.neutron [-] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.705774] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c99a267-d295-4480-9329-c32f17c4aaa7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.721207] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b5d788-8428-4528-b8ac-b82fd8f735aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.765337] env[62952]: DEBUG nova.compute.manager [req-db1c79ff-1e34-493e-9044-1eb81c1141fd req-3cb72ae8-7537-4390-af77-cb8237d65eb2 service nova] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Detach interface failed, port_id=df27cdc8-1bd7-46ad-9cfe-38ab3802d917, reason: Instance 44f0943d-7267-4472-8fe2-445a0727654a could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1080.841807] env[62952]: DEBUG nova.compute.manager [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1080.970913] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367616, 'name': ReconfigVM_Task, 'duration_secs': 0.437165} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.974427] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1081.165979] env[62952]: INFO nova.compute.manager [-] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Took 1.24 seconds to deallocate network for instance. [ 1081.222561] env[62952]: DEBUG nova.network.neutron [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [{"id": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "address": "fa:16:3e:4d:ab:95", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b40f3e8-72", "ovs_interfaceid": "4b40f3e8-7283-4edc-8016-d2a0bde62f60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.366544] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.366831] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.482538] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.482765] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.482924] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.483125] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.483279] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.483430] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.483636] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.483799] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.483966] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.484728] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.484944] env[62952]: DEBUG nova.virt.hardware [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.493095] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1081.493474] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90465cd9-4c2d-4f28-b244-e489a7fcd8c2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.521075] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1081.521075] env[62952]: value = "task-1367617" [ 1081.521075] env[62952]: _type = "Task" [ 1081.521075] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.529434] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.672552] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.726284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-867ffed2-50ec-45d8-b64e-989cf8d1b0e2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.875116] env[62952]: INFO nova.compute.claims [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.032340] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367617, 'name': ReconfigVM_Task, 'duration_secs': 0.34397} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.032637] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1082.033441] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64d9050-5c09-4265-8a80-8989ff8d3272 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.056173] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2/ea621ff2-e54e-4403-9e6d-dea84c7e8ad2.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.058309] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe400c95-add9-405d-8574-8c66e0bf5823 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.079203] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1082.079203] env[62952]: value = "task-1367618" [ 1082.079203] env[62952]: _type = "Task" [ 1082.079203] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.091389] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.222049] env[62952]: DEBUG nova.network.neutron [-] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.230688] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.231010] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19b6af10-d6c7-4cc3-b871-e1b147267868 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.235996] env[62952]: DEBUG nova.compute.manager [req-46b66261-db2f-405b-b6dc-db1896ea0f4f req-ebd1e214-7153-4088-bf94-4cc315c2f22e service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Received event network-vif-deleted-aa4656a7-42d8-47ba-a29d-817000a4b596 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.236982] env[62952]: INFO nova.compute.manager [req-46b66261-db2f-405b-b6dc-db1896ea0f4f req-ebd1e214-7153-4088-bf94-4cc315c2f22e service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Neutron deleted interface aa4656a7-42d8-47ba-a29d-817000a4b596; detaching it from the instance and deleting it from the info cache [ 1082.236982] env[62952]: DEBUG nova.network.neutron [req-46b66261-db2f-405b-b6dc-db1896ea0f4f req-ebd1e214-7153-4088-bf94-4cc315c2f22e service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.241490] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1082.241490] env[62952]: value = "task-1367619" [ 1082.241490] env[62952]: _type = "Task" [ 1082.241490] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.995470] env[62952]: INFO nova.compute.resource_tracker [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating resource usage from migration c7775c81-116a-43e5-8781-e6537d043d8b [ 1083.002750] env[62952]: INFO nova.compute.manager [-] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Took 2.43 seconds to deallocate network for instance. [ 1083.002750] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.002750] env[62952]: WARNING oslo_vmware.common.loopingcall [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] task run outlasted interval by 0.25796699999999995 sec [ 1083.005066] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e348a786-d66c-46ac-81a0-cb56a4e4c23a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.020063] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367618, 'name': ReconfigVM_Task, 'duration_secs': 0.869547} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.024238] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Reconfigured VM instance instance-00000060 to attach disk [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2/ea621ff2-e54e-4403-9e6d-dea84c7e8ad2.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.024882] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.028144] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.033767] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4611c5-67ee-44da-a499-7fc3c4595099 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.064273] env[62952]: DEBUG nova.compute.manager [req-46b66261-db2f-405b-b6dc-db1896ea0f4f req-ebd1e214-7153-4088-bf94-4cc315c2f22e service nova] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Detach interface failed, port_id=aa4656a7-42d8-47ba-a29d-817000a4b596, reason: Instance 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1083.193582] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d706ac-90da-4e77-bbdb-4d069522ebe0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.202565] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e111bac1-003c-4317-a304-cbb877ee341b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.234742] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d646a0-64c1-45f4-9d0c-2f8de945eb71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.242796] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e3db91-1ee6-4c32-b0e5-6406e0236ab7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.256382] env[62952]: DEBUG nova.compute.provider_tree [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.517579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.518240] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.530724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f452ac-6f04-4510-8239-ebf4a1801496 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.548951] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8ebfa0-b8b6-48ef-8624-3066fc54cd7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.565895] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.760062] env[62952]: DEBUG nova.scheduler.client.report [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.017112] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.104571] env[62952]: DEBUG nova.network.neutron [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Port 6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1084.265113] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.898s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.265415] env[62952]: INFO nova.compute.manager [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Migrating [ 1084.274299] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.602s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.274652] env[62952]: DEBUG nova.objects.instance [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid 44f0943d-7267-4472-8fe2-445a0727654a {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.517357] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.787195] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.787412] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.787627] env[62952]: DEBUG nova.network.neutron [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.932959] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca01c2f-3c66-4279-9678-11fb92e3f877 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.940620] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e129450c-7701-4818-8c93-cde8bca247c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.970522] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edb73d4-ea5d-4ffe-a68d-b7b5be4a5e30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.977582] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7641298f-f0b4-46f7-8de8-57119a72d1f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.991446] env[62952]: DEBUG nova.compute.provider_tree [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.016219] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.126169] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.126396] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.126571] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.494754] env[62952]: DEBUG nova.scheduler.client.report [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.508821] env[62952]: DEBUG nova.network.neutron [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance_info_cache with network_info: [{"id": "69251416-59fd-45c5-a937-fe98301eb5a7", "address": "fa:16:3e:f6:6b:d5", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69251416-59", "ovs_interfaceid": "69251416-59fd-45c5-a937-fe98301eb5a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.520101] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.000078] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.726s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.002381] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.485s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.002624] env[62952]: DEBUG nova.objects.instance [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'resources' on Instance uuid 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.011171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.021622] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.022552] env[62952]: INFO nova.scheduler.client.report [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance 44f0943d-7267-4472-8fe2-445a0727654a [ 1086.160791] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.161042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.161292] env[62952]: DEBUG nova.network.neutron [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1086.525652] env[62952]: DEBUG oslo_vmware.api [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367619, 'name': PowerOnVM_Task, 'duration_secs': 3.889246} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.526113] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.526309] env[62952]: DEBUG nova.compute.manager [None req-c55a75e5-288f-48dd-913f-e1790305a8a2 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.528835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-62b5da39-6b76-48bc-890f-a17097c0f5ad tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "44f0943d-7267-4472-8fe2-445a0727654a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.732s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.530541] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52ba779-8e10-4bd7-90bd-b5b9e3e34779 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.658279] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07e4f8e-e484-4b53-8035-d49a4a6f9eca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.667464] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e621d14a-f0a0-45da-b20d-9b0b69fc610b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.698820] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de07177-328a-4f85-8e8b-05522dda26d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.706154] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803d6f90-d982-4386-b31c-1dac40d7dbdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.719127] env[62952]: DEBUG nova.compute.provider_tree [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.854678] env[62952]: DEBUG nova.network.neutron [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [{"id": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "address": "fa:16:3e:74:a7:d4", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bbebc32-dc", "ovs_interfaceid": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.220317] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.220714] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.225061] env[62952]: DEBUG nova.scheduler.client.report [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.357243] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.530134] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d589f4e-8dfa-4918-b154-2c4f74d7894a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.549031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.549306] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.549511] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.549693] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.549863] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.551410] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.554630] env[62952]: INFO nova.compute.manager [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Terminating instance [ 1087.556299] env[62952]: DEBUG nova.compute.manager [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1087.556492] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1087.557245] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59f1479-7ddd-4c56-bfb0-5ddc3319f3f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.564018] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.564256] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-268fbb6d-98ee-43c9-984a-2154346c1493 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.569834] env[62952]: DEBUG oslo_vmware.api [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1087.569834] env[62952]: value = "task-1367620" [ 1087.569834] env[62952]: _type = "Task" [ 1087.569834] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.577635] env[62952]: DEBUG oslo_vmware.api [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.731040] env[62952]: DEBUG nova.compute.utils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1087.732930] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.758795] env[62952]: INFO nova.scheduler.client.report [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Deleted allocations for instance 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b [ 1087.877849] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e091e4-50e6-46d5-9f84-44153e9c6de7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.898169] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484c3190-60bb-4d62-af56-3a7783684280 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.905121] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1088.058086] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.058429] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01fefe1b-3b73-495d-a7fa-fef705d59ee5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.065305] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1088.065305] env[62952]: value = "task-1367621" [ 1088.065305] env[62952]: _type = "Task" [ 1088.065305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.074113] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.081751] env[62952]: DEBUG oslo_vmware.api [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367620, 'name': PowerOffVM_Task, 'duration_secs': 0.216967} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.082084] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.082304] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1088.082572] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-587a71ad-444d-46af-acf1-cf3e97fd6631 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.180227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.180227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.180227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.180500] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.180892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.184649] env[62952]: INFO nova.compute.manager [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Terminating instance [ 1088.187611] env[62952]: DEBUG nova.compute.manager [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1088.187898] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.189183] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e72ee29-e68a-4446-a3c1-5f6cea740667 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.200125] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.200493] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac2f6f97-cb9a-45de-aca0-927f0b327dfb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.207950] env[62952]: DEBUG oslo_vmware.api [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1088.207950] env[62952]: value = "task-1367623" [ 1088.207950] env[62952]: _type = "Task" [ 1088.207950] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.219550] env[62952]: DEBUG oslo_vmware.api [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367623, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.236684] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.016s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.267084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3b66f3d0-e0e6-4c15-8f16-da1466e13d90 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.822s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.411721] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.412083] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42cbe82c-bc42-4c03-8e97-8bba0b1dbb90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.420231] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1088.420231] env[62952]: value = "task-1367624" [ 1088.420231] env[62952]: _type = "Task" [ 1088.420231] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.433831] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.574991] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367621, 'name': PowerOffVM_Task, 'duration_secs': 0.384854} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.575330] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.575462] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1088.717146] env[62952]: DEBUG oslo_vmware.api [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367623, 'name': PowerOffVM_Task, 'duration_secs': 0.247389} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.718185] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.718368] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1088.718627] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1088.718809] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1088.718989] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore2] 12365f25-7d85-48dc-b2b7-9c8d0695e4a9 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.719239] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb2e61b3-d2b3-40ec-8478-ac08750cf872 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.720684] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c2dc14c-27e1-4cbd-bfea-19084ed63333 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.727114] env[62952]: DEBUG oslo_vmware.api [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1088.727114] env[62952]: value = "task-1367625" [ 1088.727114] env[62952]: _type = "Task" [ 1088.727114] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.734386] env[62952]: DEBUG oslo_vmware.api [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367625, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.799433] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1088.799711] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1088.799915] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleting the datastore file [datastore1] 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.800255] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30e89a15-f91c-47a2-8801-32a18ea36347 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.810440] env[62952]: DEBUG oslo_vmware.api [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1088.810440] env[62952]: value = "task-1367627" [ 1088.810440] env[62952]: _type = "Task" [ 1088.810440] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.818942] env[62952]: DEBUG oslo_vmware.api [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367627, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.930942] env[62952]: DEBUG oslo_vmware.api [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367624, 'name': PowerOnVM_Task, 'duration_secs': 0.389489} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.931233] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.931425] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5f1bca81-2a66-4fa4-8303-9c81731af70f tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance 'ea621ff2-e54e-4403-9e6d-dea84c7e8ad2' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.081211] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.081485] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.081683] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.081874] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.082037] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.082195] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.082408] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.082571] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.082745] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.082910] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.083099] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.088079] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-159a33f7-b527-4c3a-9bc3-19de9fab1088 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.103828] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1089.103828] env[62952]: value = "task-1367628" [ 1089.103828] env[62952]: _type = "Task" [ 1089.103828] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.111693] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367628, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.241410] env[62952]: DEBUG oslo_vmware.api [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367625, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144175} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.241825] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.242171] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.242478] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.242742] env[62952]: INFO nova.compute.manager [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Took 1.69 seconds to destroy the instance on the hypervisor. [ 1089.243104] env[62952]: DEBUG oslo.service.loopingcall [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.243446] env[62952]: DEBUG nova.compute.manager [-] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.243497] env[62952]: DEBUG nova.network.neutron [-] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.309773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.310086] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.310343] env[62952]: INFO nova.compute.manager [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Attaching volume 3aaf2335-5c53-4729-abe8-cf69d9aad29a to /dev/sdb [ 1089.324085] env[62952]: DEBUG oslo_vmware.api [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367627, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153622} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.324574] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.324789] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.324969] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.325156] env[62952]: INFO nova.compute.manager [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1089.325398] env[62952]: DEBUG oslo.service.loopingcall [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.325591] env[62952]: DEBUG nova.compute.manager [-] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.325680] env[62952]: DEBUG nova.network.neutron [-] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.353235] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e939d0-f948-483d-8e10-8434656059a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.363203] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14423d4-0973-413f-a366-19a758626ba5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.379575] env[62952]: DEBUG nova.virt.block_device [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating existing volume attachment record: 34458ce0-12e6-4e85-a107-607f88e9b57e {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1089.621110] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367628, 'name': ReconfigVM_Task, 'duration_secs': 0.196782} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.621560] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.637424] env[62952]: DEBUG nova.compute.manager [req-28ddfb01-be0a-46dd-9c8e-7722cce1711c req-72453fad-e80a-4036-abb2-77523c3cb063 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Received event network-vif-deleted-a0a6224c-fa50-4b7d-a8cd-8cf619077179 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1089.637515] env[62952]: INFO nova.compute.manager [req-28ddfb01-be0a-46dd-9c8e-7722cce1711c req-72453fad-e80a-4036-abb2-77523c3cb063 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Neutron deleted interface a0a6224c-fa50-4b7d-a8cd-8cf619077179; detaching it from the instance and deleting it from the info cache [ 1089.637701] env[62952]: DEBUG nova.network.neutron [req-28ddfb01-be0a-46dd-9c8e-7722cce1711c req-72453fad-e80a-4036-abb2-77523c3cb063 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.101482] env[62952]: DEBUG nova.network.neutron [-] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.128709] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.129022] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.129202] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.129394] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.129549] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.129739] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.129953] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.130130] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.130301] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.130470] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.130651] env[62952]: DEBUG nova.virt.hardware [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.136112] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1090.137064] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f7c3c67-d261-4882-8017-7915400bc144 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.150351] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d85acc2-e890-4982-91f7-53c9f2a16649 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.161198] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99083722-a746-4623-881e-c2246b87e37c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.173079] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1090.173079] env[62952]: value = "task-1367632" [ 1090.173079] env[62952]: _type = "Task" [ 1090.173079] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.181749] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.190094] env[62952]: DEBUG nova.compute.manager [req-28ddfb01-be0a-46dd-9c8e-7722cce1711c req-72453fad-e80a-4036-abb2-77523c3cb063 service nova] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Detach interface failed, port_id=a0a6224c-fa50-4b7d-a8cd-8cf619077179, reason: Instance 12365f25-7d85-48dc-b2b7-9c8d0695e4a9 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1090.330619] env[62952]: DEBUG nova.network.neutron [-] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.603856] env[62952]: INFO nova.compute.manager [-] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Took 1.36 seconds to deallocate network for instance. [ 1090.682923] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367632, 'name': ReconfigVM_Task, 'duration_secs': 0.370441} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.683261] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.684089] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044d7094-5d64-446b-aaa7-296ba69682b6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.706596] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed/9c18c8c7-bc97-4a84-a86d-e23f8cce20ed.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1090.706813] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8eb92059-15d3-4eb0-8ec7-61527b6107e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.724245] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1090.724245] env[62952]: value = "task-1367633" [ 1090.724245] env[62952]: _type = "Task" [ 1090.724245] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.731363] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.833682] env[62952]: INFO nova.compute.manager [-] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Took 1.51 seconds to deallocate network for instance. [ 1090.940042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.940280] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.111026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.111026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.111244] env[62952]: DEBUG nova.objects.instance [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid 12365f25-7d85-48dc-b2b7-9c8d0695e4a9 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.233772] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367633, 'name': ReconfigVM_Task, 'duration_secs': 0.333113} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.234057] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed/9c18c8c7-bc97-4a84-a86d-e23f8cce20ed.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1091.234340] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.340519] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.443321] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1091.506599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.506874] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.507081] env[62952]: DEBUG nova.compute.manager [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Going to confirm migration 3 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1091.666351] env[62952]: DEBUG nova.compute.manager [req-f89c8f8a-8057-4bee-b510-af56c612abee req-52a3a247-3565-4b44-a98c-bcab17d4c90c service nova] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Received event network-vif-deleted-4b40f3e8-7283-4edc-8016-d2a0bde62f60 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.740559] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef02f0f-2f13-42e0-870d-23d3b8b311d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.764821] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc132508-68f9-4b74-8c63-276eeb3542d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.782234] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.799574] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde9b5a3-8bed-4851-b865-e24129af2c34 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.806867] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b162c73-d026-47e0-ad22-2efc4f9f5b24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.839181] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37cd4dc-b69d-4ae0-8034-b332fe24193a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.847069] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b2c3bb-8c77-49b0-aa42-c38e04a5cac3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.860833] env[62952]: DEBUG nova.compute.provider_tree [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.961275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.068126] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.068334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.068518] env[62952]: DEBUG nova.network.neutron [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.068708] env[62952]: DEBUG nova.objects.instance [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'info_cache' on Instance uuid ea621ff2-e54e-4403-9e6d-dea84c7e8ad2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.324448] env[62952]: DEBUG nova.network.neutron [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Port 69251416-59fd-45c5-a937-fe98301eb5a7 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1092.364236] env[62952]: DEBUG nova.scheduler.client.report [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.869573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.873562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.533s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.873973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.876393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.915s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.878473] env[62952]: INFO nova.compute.claims [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.891838] env[62952]: INFO nova.scheduler.client.report [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance 12365f25-7d85-48dc-b2b7-9c8d0695e4a9 [ 1092.898925] env[62952]: INFO nova.scheduler.client.report [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted allocations for instance 867ffed2-50ec-45d8-b64e-989cf8d1b0e2 [ 1093.285975] env[62952]: DEBUG nova.network.neutron [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [{"id": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "address": "fa:16:3e:74:a7:d4", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bbebc32-dc", "ovs_interfaceid": "6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.347306] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.347540] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.347719] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.398919] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9b61f722-9988-4aae-8864-b25ef4c6b4a1 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "12365f25-7d85-48dc-b2b7-9c8d0695e4a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.849s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.405811] env[62952]: DEBUG oslo_concurrency.lockutils [None req-33ed4739-62ce-4578-82b3-45df52db9d10 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "867ffed2-50ec-45d8-b64e-989cf8d1b0e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.226s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.788641] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.788996] env[62952]: DEBUG nova.objects.instance [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'migration_context' on Instance uuid ea621ff2-e54e-4403-9e6d-dea84c7e8ad2 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.929716] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1093.929967] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1093.931244] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5406ac-4935-4f58-9f67-4503efa7c3e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.951624] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d46722e-84ef-41c6-87b1-1d82aeb9960f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.976446] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a/volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.979144] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68f357d5-0f8a-4d24-b0a8-06567019a64e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.998239] env[62952]: DEBUG oslo_vmware.api [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1093.998239] env[62952]: value = "task-1367635" [ 1093.998239] env[62952]: _type = "Task" [ 1093.998239] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.009042] env[62952]: DEBUG oslo_vmware.api [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367635, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.058067] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bd6006-a53f-4458-bfc8-f8099f7328ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.065508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5f1527-e1a9-4053-8abe-1f05eb60bc09 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.097302] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7526bac-04c1-46b2-8330-9f931d70e12a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.105969] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8fee7c-e558-43ef-9a51-8388f1817620 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.120517] env[62952]: DEBUG nova.compute.provider_tree [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.267870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.268179] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.291675] env[62952]: DEBUG nova.objects.base [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1094.292583] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fa3af1-3618-4f8b-943e-96a30d1c01f4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.311258] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3639b2e7-f3b8-48e9-9349-2f9da962196b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.316782] env[62952]: DEBUG oslo_vmware.api [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1094.316782] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52517e58-f092-1b41-6e22-4cde83acbbcb" [ 1094.316782] env[62952]: _type = "Task" [ 1094.316782] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.324123] env[62952]: DEBUG oslo_vmware.api [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52517e58-f092-1b41-6e22-4cde83acbbcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.396502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.396748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.396960] env[62952]: DEBUG nova.network.neutron [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.507316] env[62952]: DEBUG oslo_vmware.api [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367635, 'name': ReconfigVM_Task, 'duration_secs': 0.354827} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.507630] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a/volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.512423] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fcbb568-9990-4c3b-b0f3-2075dce6e6c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.527512] env[62952]: DEBUG oslo_vmware.api [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1094.527512] env[62952]: value = "task-1367636" [ 1094.527512] env[62952]: _type = "Task" [ 1094.527512] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.538765] env[62952]: DEBUG oslo_vmware.api [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367636, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.585502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "f302233b-2e1f-4c71-b000-0c9eb6530730" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.585749] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.624065] env[62952]: DEBUG nova.scheduler.client.report [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.770962] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1094.828175] env[62952]: DEBUG oslo_vmware.api [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52517e58-f092-1b41-6e22-4cde83acbbcb, 'name': SearchDatastore_Task, 'duration_secs': 0.008994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.828536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.038359] env[62952]: DEBUG oslo_vmware.api [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367636, 'name': ReconfigVM_Task, 'duration_secs': 0.134993} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.038713] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1095.087632] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1095.110946] env[62952]: DEBUG nova.network.neutron [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance_info_cache with network_info: [{"id": "69251416-59fd-45c5-a937-fe98301eb5a7", "address": "fa:16:3e:f6:6b:d5", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69251416-59", "ovs_interfaceid": "69251416-59fd-45c5-a937-fe98301eb5a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.128243] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.252s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.128744] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1095.131454] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.303s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.292393] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.605946] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.614162] env[62952]: DEBUG oslo_concurrency.lockutils [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.634422] env[62952]: DEBUG nova.compute.utils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1095.638607] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1095.638829] env[62952]: DEBUG nova.network.neutron [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1095.677103] env[62952]: DEBUG nova.policy [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b4bcad9f5ed4ee7882df428957d1d93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8427597d4aae451e9375ade7c06b2bcf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1095.795754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbd04bb-4c2d-470b-a856-ee3fb3831b3a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.804874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a95df9-2641-4d3e-8f13-c3a63f94c004 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.836911] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98b948b-c8ac-4ca7-b69a-438236bee2bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.844507] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b34c867-c477-47a9-a439-5b28562c6efe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.858614] env[62952]: DEBUG nova.compute.provider_tree [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.015208] env[62952]: DEBUG nova.network.neutron [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Successfully created port: 2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1096.074871] env[62952]: DEBUG nova.objects.instance [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'flavor' on Instance uuid f5e5c5c1-e2af-40e0-a957-67b0fd22718d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.140430] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1096.143581] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de039a30-33a4-406d-a2f5-ea6853a3d520 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.162963] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af8ae42-691a-4150-9b47-2d35b34b8b2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.171018] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1096.361521] env[62952]: DEBUG nova.scheduler.client.report [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1096.579724] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0aa6aa9f-df83-4265-b0a7-94b0afdc0f39 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.269s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.675543] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1096.675858] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e73edb07-0c79-4695-83f8-90a764b4533c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.683400] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1096.683400] env[62952]: value = "task-1367637" [ 1096.683400] env[62952]: _type = "Task" [ 1096.683400] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.692522] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367637, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.934775] env[62952]: INFO nova.compute.manager [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Rebuilding instance [ 1096.978210] env[62952]: DEBUG nova.compute.manager [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.979132] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839a28e1-9c21-439c-b91b-000061aaa810 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.154889] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1097.174823] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1097.175100] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1097.175837] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1097.175837] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1097.175837] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1097.175837] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1097.176110] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1097.176200] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1097.176380] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1097.176547] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1097.176725] env[62952]: DEBUG nova.virt.hardware [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1097.177596] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17da359-8353-4cc1-8228-a9659135bf8d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.188054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44de6d33-871b-4ef1-9a57-f23ce576eb59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.196592] env[62952]: DEBUG oslo_vmware.api [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367637, 'name': PowerOnVM_Task, 'duration_secs': 0.458303} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.204282] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1097.204480] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-88d64ffd-11d2-456c-8cf4-83409daeb08e tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance '9c18c8c7-bc97-4a84-a86d-e23f8cce20ed' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1097.372398] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.241s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.375258] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.083s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.377416] env[62952]: INFO nova.compute.claims [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.423103] env[62952]: DEBUG nova.compute.manager [req-02e39fcc-a86f-46df-8b60-c273405aa1c6 req-f473f7c5-ef97-4623-bfa0-acaf9834839c service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Received event network-vif-plugged-2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.423351] env[62952]: DEBUG oslo_concurrency.lockutils [req-02e39fcc-a86f-46df-8b60-c273405aa1c6 req-f473f7c5-ef97-4623-bfa0-acaf9834839c service nova] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.423712] env[62952]: DEBUG oslo_concurrency.lockutils [req-02e39fcc-a86f-46df-8b60-c273405aa1c6 req-f473f7c5-ef97-4623-bfa0-acaf9834839c service nova] Lock "287250a7-3354-49b7-9194-d88ae51afdaf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.423750] env[62952]: DEBUG oslo_concurrency.lockutils [req-02e39fcc-a86f-46df-8b60-c273405aa1c6 req-f473f7c5-ef97-4623-bfa0-acaf9834839c service nova] Lock "287250a7-3354-49b7-9194-d88ae51afdaf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.423958] env[62952]: DEBUG nova.compute.manager [req-02e39fcc-a86f-46df-8b60-c273405aa1c6 req-f473f7c5-ef97-4623-bfa0-acaf9834839c service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] No waiting events found dispatching network-vif-plugged-2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1097.424127] env[62952]: WARNING nova.compute.manager [req-02e39fcc-a86f-46df-8b60-c273405aa1c6 req-f473f7c5-ef97-4623-bfa0-acaf9834839c service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Received unexpected event network-vif-plugged-2595a3f5-136a-4d28-83ff-9bcc6d05154e for instance with vm_state building and task_state spawning. [ 1097.490044] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.490609] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2bab9afc-d572-4af8-93d7-ea616b8d9fad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.498806] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1097.498806] env[62952]: value = "task-1367638" [ 1097.498806] env[62952]: _type = "Task" [ 1097.498806] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.507486] env[62952]: DEBUG nova.network.neutron [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Successfully updated port: 2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1097.515577] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.935167] env[62952]: INFO nova.scheduler.client.report [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocation for migration 3effe577-bf29-41de-a612-7938583e828a [ 1098.008070] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367638, 'name': PowerOffVM_Task, 'duration_secs': 0.213856} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.008605] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.012031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.012031] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.012031] env[62952]: DEBUG nova.network.neutron [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1098.068354] env[62952]: INFO nova.compute.manager [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Detaching volume 3aaf2335-5c53-4729-abe8-cf69d9aad29a [ 1098.101509] env[62952]: INFO nova.virt.block_device [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Attempting to driver detach volume 3aaf2335-5c53-4729-abe8-cf69d9aad29a from mountpoint /dev/sdb [ 1098.101758] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1098.101970] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1098.102858] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4165cc-d267-4056-b20c-62d57ca9ac12 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.124459] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fde4888-93e8-46bd-8b30-3e4893a3d20a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.131673] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452deea7-f65a-4faf-a7bf-2e888188f4aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.152318] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a040947-5483-4aca-894f-063645996362 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.170883] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] The volume has not been displaced from its original location: [datastore2] volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a/volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1098.176090] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1098.176994] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97560cd4-4af0-4747-ba6a-abf1dd5c2e58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.194716] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1098.194716] env[62952]: value = "task-1367639" [ 1098.194716] env[62952]: _type = "Task" [ 1098.194716] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.202291] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367639, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.442235] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6398a078-aaf8-4aef-869f-709a1cd045f0 tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.935s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.528022] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad2528d-0cf9-4a6a-b3ed-d38b532fb8f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.535869] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05de3409-f44c-4b00-9e4c-6a49447b0ac8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.565621] env[62952]: DEBUG nova.network.neutron [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1098.568168] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67db528e-5076-4e86-ada6-3d94fb491274 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.576696] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f077697a-0295-404e-a7fc-41bb512b12e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.590658] env[62952]: DEBUG nova.compute.provider_tree [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.658253] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.658579] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.658813] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.659023] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.659203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.661380] env[62952]: INFO nova.compute.manager [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Terminating instance [ 1098.663175] env[62952]: DEBUG nova.compute.manager [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1098.663369] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1098.664217] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179883f9-820b-4b83-ae63-652de7d3f2ad {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.673652] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.673908] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0283d02d-569a-431d-bf7b-47a7f6a9c1af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.679771] env[62952]: DEBUG oslo_vmware.api [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1098.679771] env[62952]: value = "task-1367640" [ 1098.679771] env[62952]: _type = "Task" [ 1098.679771] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.687254] env[62952]: DEBUG oslo_vmware.api [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367640, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.703430] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367639, 'name': ReconfigVM_Task, 'duration_secs': 0.225879} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.704398] env[62952]: DEBUG nova.network.neutron [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updating instance_info_cache with network_info: [{"id": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "address": "fa:16:3e:b0:9b:f1", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2595a3f5-13", "ovs_interfaceid": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.705597] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1098.711075] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-756e9935-6345-4b01-b469-ba99416d4b0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.733425] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1098.733425] env[62952]: value = "task-1367641" [ 1098.733425] env[62952]: _type = "Task" [ 1098.733425] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.742362] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367641, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.071079] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "9036f863-9d77-491b-8aa9-978d0c400708" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.071339] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "9036f863-9d77-491b-8aa9-978d0c400708" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.095266] env[62952]: DEBUG nova.scheduler.client.report [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.189655] env[62952]: DEBUG oslo_vmware.api [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367640, 'name': PowerOffVM_Task, 'duration_secs': 0.22441} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.190038] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.190099] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.190348] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-877afbbc-e70b-4d76-b44f-d2fad6ca0751 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.211163] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.211419] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Instance network_info: |[{"id": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "address": "fa:16:3e:b0:9b:f1", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2595a3f5-13", "ovs_interfaceid": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1099.211856] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:9b:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2595a3f5-136a-4d28-83ff-9bcc6d05154e', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1099.219860] env[62952]: DEBUG oslo.service.loopingcall [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.220109] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1099.220349] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c87a0ecc-26ca-4100-9764-3bf1f42e9d54 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.243583] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367641, 'name': ReconfigVM_Task, 'duration_secs': 0.177313} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.244992] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1099.247419] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1099.247419] env[62952]: value = "task-1367643" [ 1099.247419] env[62952]: _type = "Task" [ 1099.247419] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.255182] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367643, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.451486] env[62952]: DEBUG nova.compute.manager [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Received event network-changed-2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.451739] env[62952]: DEBUG nova.compute.manager [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Refreshing instance network info cache due to event network-changed-2595a3f5-136a-4d28-83ff-9bcc6d05154e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1099.451996] env[62952]: DEBUG oslo_concurrency.lockutils [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] Acquiring lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.452340] env[62952]: DEBUG oslo_concurrency.lockutils [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] Acquired lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.452543] env[62952]: DEBUG nova.network.neutron [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Refreshing network info cache for port 2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1099.544107] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.544620] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.544933] env[62952]: DEBUG nova.compute.manager [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Going to confirm migration 4 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1099.573435] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1099.600215] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.600825] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1099.604794] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.999s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.606009] env[62952]: INFO nova.compute.claims [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1099.757720] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367643, 'name': CreateVM_Task, 'duration_secs': 0.311881} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.757903] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1099.758654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.758835] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.759287] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1099.759576] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7158e08a-0de8-4a6d-974b-723b927dbdbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.764655] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1099.764655] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]527cfcc0-2bd7-2b4d-9553-df4fe475c10c" [ 1099.764655] env[62952]: _type = "Task" [ 1099.764655] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.773993] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527cfcc0-2bd7-2b4d-9553-df4fe475c10c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.092759] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.093063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquired lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.093270] env[62952]: DEBUG nova.network.neutron [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1100.093464] env[62952]: DEBUG nova.objects.instance [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lazy-loading 'info_cache' on Instance uuid 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.097438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.113758] env[62952]: DEBUG nova.compute.utils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1100.116872] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1100.116872] env[62952]: DEBUG nova.network.neutron [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1100.174079] env[62952]: DEBUG nova.policy [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8171575c58594e7ebc37af7c4c858c10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c451895447b473eb18675f6bc843081', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1100.233464] env[62952]: DEBUG nova.network.neutron [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updated VIF entry in instance network info cache for port 2595a3f5-136a-4d28-83ff-9bcc6d05154e. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1100.233844] env[62952]: DEBUG nova.network.neutron [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updating instance_info_cache with network_info: [{"id": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "address": "fa:16:3e:b0:9b:f1", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2595a3f5-13", "ovs_interfaceid": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.274543] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527cfcc0-2bd7-2b4d-9553-df4fe475c10c, 'name': SearchDatastore_Task, 'duration_secs': 0.010458} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.274846] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.275102] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1100.275347] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.275499] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.275680] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1100.276196] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a4d8eb5-7630-48d4-88b2-28947dbd04a9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.284550] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1100.284704] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1100.285402] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8b522b0-e82f-413f-b448-8037566f38b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.290136] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1100.290136] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52924a60-5725-8826-ea6d-e749afabc1bf" [ 1100.290136] env[62952]: _type = "Task" [ 1100.290136] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.297287] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52924a60-5725-8826-ea6d-e749afabc1bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.297756] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.298013] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6ede426-fe3d-4ce9-9413-8fc1d7e60c02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.303963] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1100.303963] env[62952]: value = "task-1367644" [ 1100.303963] env[62952]: _type = "Task" [ 1100.303963] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.311659] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367644, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.319930] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1100.320147] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1100.320326] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore1] ea621ff2-e54e-4403-9e6d-dea84c7e8ad2 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.320587] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-138440ab-deda-4738-b1fa-dd0dc9af648f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.327250] env[62952]: DEBUG oslo_vmware.api [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1100.327250] env[62952]: value = "task-1367645" [ 1100.327250] env[62952]: _type = "Task" [ 1100.327250] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.334495] env[62952]: DEBUG oslo_vmware.api [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367645, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.454426] env[62952]: DEBUG nova.network.neutron [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Successfully created port: 612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1100.617552] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1100.736712] env[62952]: DEBUG oslo_concurrency.lockutils [req-1774d0f0-d63b-4ace-b30d-6261c0188b1b req-40a0e026-2cd1-4e1f-8118-31664270cf65 service nova] Releasing lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.770074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a234605-a792-4880-8542-af08873b5291 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.777265] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b900fe9-0e62-44d5-b5ed-80da80096d94 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.814625] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce227258-eda1-469c-8cfb-3d72e43257e5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.822901] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52924a60-5725-8826-ea6d-e749afabc1bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009353} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.828949] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96afc40a-5b9d-42ba-b094-ddad74ee0d58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.832118] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbb05e9-371c-4781-a99d-bbf222dead6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.837980] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1100.838206] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1100.838401] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1100.839088] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447ee91a-a018-4ac8-b0e0-22fa81d4924c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.844291] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1100.844291] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]520d46be-9f09-01d0-a56c-a02d10de9656" [ 1100.844291] env[62952]: _type = "Task" [ 1100.844291] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.867178] env[62952]: DEBUG oslo_vmware.api [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367645, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189953} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.867654] env[62952]: DEBUG nova.compute.provider_tree [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.871506] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.871699] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1100.871874] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1100.872065] env[62952]: INFO nova.compute.manager [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Took 2.21 seconds to destroy the instance on the hypervisor. [ 1100.872300] env[62952]: DEBUG oslo.service.loopingcall [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.872963] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3ad585-053e-4866-b22e-6d93b153e0a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.875400] env[62952]: DEBUG nova.compute.manager [-] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1100.875498] env[62952]: DEBUG nova.network.neutron [-] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1100.883126] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520d46be-9f09-01d0-a56c-a02d10de9656, 'name': SearchDatastore_Task, 'duration_secs': 0.012085} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.884273] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.884528] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 287250a7-3354-49b7-9194-d88ae51afdaf/287250a7-3354-49b7-9194-d88ae51afdaf.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1100.885441] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1100.885733] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1100.886008] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff49067a-8cc4-4e81-9962-77644c23eceb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.888315] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904654bd-b141-46d1-a53c-9719453813f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.897108] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1100.897108] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6566919-716f-4132-ae1d-69d95323abc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.898128] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1100.898128] env[62952]: value = "task-1367646" [ 1100.898128] env[62952]: _type = "Task" [ 1100.898128] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.908197] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367646, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.019097] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1101.019097] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1101.019302] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.019507] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c671b5a-6f4d-4fa0-8d3f-47c1435c2071 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.026498] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1101.026498] env[62952]: value = "task-1367648" [ 1101.026498] env[62952]: _type = "Task" [ 1101.026498] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.034583] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367648, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.376841] env[62952]: DEBUG nova.scheduler.client.report [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.408866] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367646, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477079} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.409382] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 287250a7-3354-49b7-9194-d88ae51afdaf/287250a7-3354-49b7-9194-d88ae51afdaf.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1101.409382] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1101.409599] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-981b38dd-6203-425d-8800-5afc1eef8d4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.417171] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1101.417171] env[62952]: value = "task-1367649" [ 1101.417171] env[62952]: _type = "Task" [ 1101.417171] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.425168] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367649, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.448744] env[62952]: DEBUG nova.network.neutron [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance_info_cache with network_info: [{"id": "69251416-59fd-45c5-a937-fe98301eb5a7", "address": "fa:16:3e:f6:6b:d5", "network": {"id": "112d789e-e279-42e9-b1ee-37fb246787d2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1919589714-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "094ca56818604bee9c1f75d1493ab3af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69251416-59", "ovs_interfaceid": "69251416-59fd-45c5-a937-fe98301eb5a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.478971] env[62952]: DEBUG nova.compute.manager [req-cb0f4aad-a7b9-430b-b64f-24b28ebd77f3 req-109213a3-161c-44fa-bf1e-18b17a2b1a14 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Received event network-vif-deleted-6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.478971] env[62952]: INFO nova.compute.manager [req-cb0f4aad-a7b9-430b-b64f-24b28ebd77f3 req-109213a3-161c-44fa-bf1e-18b17a2b1a14 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Neutron deleted interface 6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36; detaching it from the instance and deleting it from the info cache [ 1101.479096] env[62952]: DEBUG nova.network.neutron [req-cb0f4aad-a7b9-430b-b64f-24b28ebd77f3 req-109213a3-161c-44fa-bf1e-18b17a2b1a14 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.536086] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367648, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226009} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.536346] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.536536] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1101.536714] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1101.615939] env[62952]: DEBUG nova.network.neutron [-] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.632009] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1101.656837] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.657137] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.657304] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.657491] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.657641] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.657791] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.657999] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.658221] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.659027] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.659027] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.659027] env[62952]: DEBUG nova.virt.hardware [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.659892] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753c74d1-a2a8-4baf-905f-913a0eefbd3b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.668305] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1717599f-c8a0-4b57-9b67-a7f23ed8dd59 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.885018] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.885018] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1101.885844] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.788s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.887318] env[62952]: INFO nova.compute.claims [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1101.909853] env[62952]: DEBUG nova.network.neutron [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Successfully updated port: 612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1101.926730] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367649, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155122} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.927870] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.928774] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de7c0b4-2178-4ec2-a35c-8eb39e2015f5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.951377] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 287250a7-3354-49b7-9194-d88ae51afdaf/287250a7-3354-49b7-9194-d88ae51afdaf.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.952132] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Releasing lock "refresh_cache-9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.952375] env[62952]: DEBUG nova.objects.instance [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lazy-loading 'migration_context' on Instance uuid 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.953427] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92976b8d-8fa3-4817-864d-db4b1b6517d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.974857] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1101.974857] env[62952]: value = "task-1367650" [ 1101.974857] env[62952]: _type = "Task" [ 1101.974857] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.983592] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367650, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.983867] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c56e87e-138a-4fbb-b32e-20c7b513a3ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.992262] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c10347c-71af-4165-9526-a7d31852ae93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.020580] env[62952]: DEBUG nova.compute.manager [req-cb0f4aad-a7b9-430b-b64f-24b28ebd77f3 req-109213a3-161c-44fa-bf1e-18b17a2b1a14 service nova] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Detach interface failed, port_id=6bbebc32-dcc0-4eb7-8d10-4e1ae25a2a36, reason: Instance ea621ff2-e54e-4403-9e6d-dea84c7e8ad2 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1102.041380] env[62952]: INFO nova.virt.block_device [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Booting with volume 3aaf2335-5c53-4729-abe8-cf69d9aad29a at /dev/sdb [ 1102.079465] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8a9e7eb-c1ff-42c8-a9f1-f735c61d2284 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.088654] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fb645c-1bd7-4fbf-9edc-f59f62efb187 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.117791] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ca9fde9-bd15-4c40-bc00-5af19daa0ddf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.119896] env[62952]: INFO nova.compute.manager [-] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Took 1.24 seconds to deallocate network for instance. [ 1102.132064] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d05f677-37fc-4c4c-99c9-7ef8158c3ab6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.158506] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6d406b-19f2-4263-ae98-db999a0a8998 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.165837] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0961a8e1-9207-4ab4-8284-99fdb61b2ce9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.180010] env[62952]: DEBUG nova.virt.block_device [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating existing volume attachment record: 490e9c46-0bdc-458e-8d4b-79f0808d2c73 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1102.395050] env[62952]: DEBUG nova.compute.utils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.396250] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1102.396440] env[62952]: DEBUG nova.network.neutron [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1102.414294] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.414457] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.414616] env[62952]: DEBUG nova.network.neutron [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.437125] env[62952]: DEBUG nova.policy [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1102.470524] env[62952]: DEBUG nova.objects.base [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Object Instance<9c18c8c7-bc97-4a84-a86d-e23f8cce20ed> lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1102.471567] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b238893a-5d0e-4b10-bac5-6fb3d72b7ba3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.494184] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ee49ca9-55a6-4e35-a58d-1fc3e9ab0d5a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.500158] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367650, 'name': ReconfigVM_Task, 'duration_secs': 0.518647} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.500781] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 287250a7-3354-49b7-9194-d88ae51afdaf/287250a7-3354-49b7-9194-d88ae51afdaf.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.501407] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10f5e30b-bf03-4047-b7dd-6fe811b77df6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.504137] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1102.504137] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52df7221-0a10-72b7-7a75-0ccf9132e946" [ 1102.504137] env[62952]: _type = "Task" [ 1102.504137] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.509044] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1102.509044] env[62952]: value = "task-1367651" [ 1102.509044] env[62952]: _type = "Task" [ 1102.509044] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.515510] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52df7221-0a10-72b7-7a75-0ccf9132e946, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.519717] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367651, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.627736] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.685149] env[62952]: DEBUG nova.network.neutron [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Successfully created port: 371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1102.901261] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1102.945553] env[62952]: DEBUG nova.network.neutron [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1103.017694] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52df7221-0a10-72b7-7a75-0ccf9132e946, 'name': SearchDatastore_Task, 'duration_secs': 0.007482} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.020130] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.026850] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367651, 'name': Rename_Task, 'duration_secs': 0.156411} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.027077] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.027331] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d14e1fd4-6eed-4ff6-97ff-7fb4bc43c6f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.034981] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1103.034981] env[62952]: value = "task-1367652" [ 1103.034981] env[62952]: _type = "Task" [ 1103.034981] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.042019] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.064730] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d608e9a7-3f0a-4a78-86fe-5cd1fa9e8cff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.071413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114f7a89-01fb-4eee-bf15-bede42a3655d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.101414] env[62952]: DEBUG nova.network.neutron [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.103081] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951c8463-b47a-402c-ad9c-fe112b51a2f0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.110276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9ad538-d272-43a4-8c51-16931ac13752 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.123799] env[62952]: DEBUG nova.compute.provider_tree [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.504277] env[62952]: DEBUG nova.compute.manager [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Received event network-vif-plugged-612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.504504] env[62952]: DEBUG oslo_concurrency.lockutils [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.504717] env[62952]: DEBUG oslo_concurrency.lockutils [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.504893] env[62952]: DEBUG oslo_concurrency.lockutils [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.505079] env[62952]: DEBUG nova.compute.manager [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] No waiting events found dispatching network-vif-plugged-612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1103.505250] env[62952]: WARNING nova.compute.manager [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Received unexpected event network-vif-plugged-612c5fea-4183-448f-9326-6ed1888b591d for instance with vm_state building and task_state spawning. [ 1103.505469] env[62952]: DEBUG nova.compute.manager [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Received event network-changed-612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.505655] env[62952]: DEBUG nova.compute.manager [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Refreshing instance network info cache due to event network-changed-612c5fea-4183-448f-9326-6ed1888b591d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1103.505828] env[62952]: DEBUG oslo_concurrency.lockutils [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.544367] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367652, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.606965] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.607345] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Instance network_info: |[{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1103.607689] env[62952]: DEBUG oslo_concurrency.lockutils [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.607840] env[62952]: DEBUG nova.network.neutron [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Refreshing network info cache for port 612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1103.609120] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:9c:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '612c5fea-4183-448f-9326-6ed1888b591d', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.616834] env[62952]: DEBUG oslo.service.loopingcall [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.617767] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.618009] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f5c92d1-14d9-43c4-83be-098fc8469842 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.633695] env[62952]: DEBUG nova.scheduler.client.report [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.642636] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.642636] env[62952]: value = "task-1367653" [ 1103.642636] env[62952]: _type = "Task" [ 1103.642636] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.651420] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367653, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.910897] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1103.937205] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.937469] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.937651] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.937890] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.938062] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.938278] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.938704] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.938796] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.939032] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.939387] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.939539] env[62952]: DEBUG nova.virt.hardware [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.941176] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9faee9b-4a27-456e-8d53-6823b50abdf1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.949083] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e468c620-0552-41df-807d-044cd24dfcfd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.044445] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367652, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.139439] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.139960] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1104.142705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.515s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.142961] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.145389] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.125s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.159353] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367653, 'name': CreateVM_Task, 'duration_secs': 0.301594} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.159587] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.160375] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.160546] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.160895] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1104.161254] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cff5e347-c463-479b-acba-904d0f82147e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.166295] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1104.166295] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5210efa4-ef2f-72d9-f1cc-b73047d45d13" [ 1104.166295] env[62952]: _type = "Task" [ 1104.166295] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.175510] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5210efa4-ef2f-72d9-f1cc-b73047d45d13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.179853] env[62952]: INFO nova.scheduler.client.report [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocations for instance ea621ff2-e54e-4403-9e6d-dea84c7e8ad2 [ 1104.291554] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.291944] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.292171] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.292385] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.292724] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.292924] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.293174] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.293352] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.293536] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.293725] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.293929] env[62952]: DEBUG nova.virt.hardware [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.294790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b88c91-43fc-4f64-b255-5e7d07e3276c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.304161] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b17a67d-5b6c-487a-9a2c-e1c8704034ef {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.320374] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:bc:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be83803d-9485-4459-882a-5dc9383e0f62', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1104.327679] env[62952]: DEBUG oslo.service.loopingcall [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.328619] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1104.328619] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c6dab3a-f7ae-455d-a1c3-558da50991c9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.342595] env[62952]: DEBUG nova.network.neutron [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updated VIF entry in instance network info cache for port 612c5fea-4183-448f-9326-6ed1888b591d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1104.342964] env[62952]: DEBUG nova.network.neutron [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.348254] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1104.348254] env[62952]: value = "task-1367654" [ 1104.348254] env[62952]: _type = "Task" [ 1104.348254] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.356760] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367654, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.382027] env[62952]: DEBUG nova.network.neutron [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Successfully updated port: 371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1104.545670] env[62952]: DEBUG oslo_vmware.api [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367652, 'name': PowerOnVM_Task, 'duration_secs': 1.027343} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.545953] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.546165] env[62952]: INFO nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Took 7.39 seconds to spawn the instance on the hypervisor. [ 1104.546345] env[62952]: DEBUG nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.547153] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33236f6a-ab3e-4408-a415-0e6b0ce16c4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.649300] env[62952]: DEBUG nova.compute.utils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1104.650743] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1104.650917] env[62952]: DEBUG nova.network.neutron [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1104.680042] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5210efa4-ef2f-72d9-f1cc-b73047d45d13, 'name': SearchDatastore_Task, 'duration_secs': 0.009899} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.680385] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.680699] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.680951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.681121] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.681307] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.681573] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa81d4bf-63e0-449d-87da-790d6fca07d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.688912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6fa5ab1d-1c8e-4b5a-b519-6ff408589fbf tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "ea621ff2-e54e-4403-9e6d-dea84c7e8ad2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.030s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.692198] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.692426] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.694329] env[62952]: DEBUG nova.policy [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17b50c91ccd347e4b45126a8f23c0278', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9657a113032417fb4b97716a772f8ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1104.696201] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2de6f62-568b-4f12-a3c4-1a71b23aa8c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.703780] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1104.703780] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5287e0da-bd1a-cbfb-906c-ad47f6998f8f" [ 1104.703780] env[62952]: _type = "Task" [ 1104.703780] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.711780] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5287e0da-bd1a-cbfb-906c-ad47f6998f8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.800524] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0beb7f-31e6-4a6d-9d2a-ed64e07e36be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.808231] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1549bfc-0b84-427f-b295-3f013512f16a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.840567] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8241a83-0202-45b7-840f-2f8fbf3a6ffc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.845414] env[62952]: DEBUG oslo_concurrency.lockutils [req-79479c29-d4f5-4f8e-97fd-09ce4508d83f req-6e467d70-d2b5-4eb2-9e99-9f6c03df14d8 service nova] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.848778] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4a0147-658f-4d61-89b2-03169de3c423 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.864925] env[62952]: DEBUG nova.compute.provider_tree [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.869699] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367654, 'name': CreateVM_Task, 'duration_secs': 0.393841} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.870008] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.870495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.870675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.870988] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1104.871248] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22809d90-2230-4c5f-9062-3cc850848b57 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.876354] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1104.876354] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52819511-f419-2c17-841e-aa3e9fa33efa" [ 1104.876354] env[62952]: _type = "Task" [ 1104.876354] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.884234] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-f302233b-2e1f-4c71-b000-0c9eb6530730" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.884378] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-f302233b-2e1f-4c71-b000-0c9eb6530730" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.884547] env[62952]: DEBUG nova.network.neutron [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.886310] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52819511-f419-2c17-841e-aa3e9fa33efa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.957228] env[62952]: DEBUG nova.network.neutron [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Successfully created port: 210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1105.063885] env[62952]: INFO nova.compute.manager [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Took 13.12 seconds to build instance. [ 1105.155649] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1105.215442] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5287e0da-bd1a-cbfb-906c-ad47f6998f8f, 'name': SearchDatastore_Task, 'duration_secs': 0.009344} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.216287] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-398e0861-5b23-409c-ad8f-97cddd5dad6f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.221548] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1105.221548] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5208050a-7413-1910-f3bd-60eceef98473" [ 1105.221548] env[62952]: _type = "Task" [ 1105.221548] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.229114] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5208050a-7413-1910-f3bd-60eceef98473, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.368693] env[62952]: DEBUG nova.scheduler.client.report [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.388675] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52819511-f419-2c17-841e-aa3e9fa33efa, 'name': SearchDatastore_Task, 'duration_secs': 0.008976} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.388976] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.389224] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1105.389438] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.416348] env[62952]: DEBUG nova.network.neutron [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1105.539996] env[62952]: DEBUG nova.network.neutron [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Updating instance_info_cache with network_info: [{"id": "371d4d90-8873-4e1a-a839-449587dadfe0", "address": "fa:16:3e:3d:77:12", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap371d4d90-88", "ovs_interfaceid": "371d4d90-8873-4e1a-a839-449587dadfe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.566156] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ef0046d-3703-49c7-bb37-0091afcf7dfd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.626s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.708719] env[62952]: DEBUG nova.compute.manager [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Received event network-vif-plugged-371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.708940] env[62952]: DEBUG oslo_concurrency.lockutils [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] Acquiring lock "f302233b-2e1f-4c71-b000-0c9eb6530730-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.709159] env[62952]: DEBUG oslo_concurrency.lockutils [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.709336] env[62952]: DEBUG oslo_concurrency.lockutils [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.709515] env[62952]: DEBUG nova.compute.manager [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] No waiting events found dispatching network-vif-plugged-371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1105.709687] env[62952]: WARNING nova.compute.manager [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Received unexpected event network-vif-plugged-371d4d90-8873-4e1a-a839-449587dadfe0 for instance with vm_state building and task_state spawning. [ 1105.709856] env[62952]: DEBUG nova.compute.manager [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Received event network-changed-371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.710031] env[62952]: DEBUG nova.compute.manager [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Refreshing instance network info cache due to event network-changed-371d4d90-8873-4e1a-a839-449587dadfe0. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1105.710209] env[62952]: DEBUG oslo_concurrency.lockutils [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] Acquiring lock "refresh_cache-f302233b-2e1f-4c71-b000-0c9eb6530730" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.733717] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5208050a-7413-1910-f3bd-60eceef98473, 'name': SearchDatastore_Task, 'duration_secs': 0.009196} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.733973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.734217] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.734502] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.734696] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1105.734920] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-feb3915e-1636-4975-8dbb-55e30a0aa0be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.736843] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-084f91e4-53ef-4411-b016-cb4d09ee803b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.745034] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1105.745034] env[62952]: value = "task-1367655" [ 1105.745034] env[62952]: _type = "Task" [ 1105.745034] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.748915] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1105.748915] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1105.749595] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c755e83e-6966-45de-8d4b-cdbc968b5fc8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.754883] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.758073] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1105.758073] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52c8ac17-5e05-8e4d-4be1-44801cc44e11" [ 1105.758073] env[62952]: _type = "Task" [ 1105.758073] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.765579] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c8ac17-5e05-8e4d-4be1-44801cc44e11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.046030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-f302233b-2e1f-4c71-b000-0c9eb6530730" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.046030] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Instance network_info: |[{"id": "371d4d90-8873-4e1a-a839-449587dadfe0", "address": "fa:16:3e:3d:77:12", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap371d4d90-88", "ovs_interfaceid": "371d4d90-8873-4e1a-a839-449587dadfe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1106.046030] env[62952]: DEBUG oslo_concurrency.lockutils [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] Acquired lock "refresh_cache-f302233b-2e1f-4c71-b000-0c9eb6530730" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.046030] env[62952]: DEBUG nova.network.neutron [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Refreshing network info cache for port 371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.046030] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:77:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '371d4d90-8873-4e1a-a839-449587dadfe0', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1106.056746] env[62952]: DEBUG oslo.service.loopingcall [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.058348] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1106.059093] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-332f3448-794e-4658-a67d-438f0279ade6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.085270] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1106.085270] env[62952]: value = "task-1367656" [ 1106.085270] env[62952]: _type = "Task" [ 1106.085270] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.096460] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367656, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.166691] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1106.199221] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.199221] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.199221] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.199709] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.200040] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.200482] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.203025] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.203025] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.203025] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.203025] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.203025] env[62952]: DEBUG nova.virt.hardware [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.203025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664455df-c456-4261-9e7f-4d648c51552f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.211313] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76095cb5-6153-4bb3-b23c-772089a2d490 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.255840] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367655, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.268980] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52c8ac17-5e05-8e4d-4be1-44801cc44e11, 'name': SearchDatastore_Task, 'duration_secs': 0.021094} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.272132] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f638a360-3220-4f5d-af1b-501013d18056 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.275715] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1106.275715] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526eef77-de94-e8c5-e439-bd761663e9fc" [ 1106.275715] env[62952]: _type = "Task" [ 1106.275715] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.285294] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526eef77-de94-e8c5-e439-bd761663e9fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.382063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.234s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.458668] env[62952]: DEBUG nova.compute.manager [req-b2ddaee9-3be3-42eb-83ba-e665a8c15394 req-6b55a99c-45c0-4788-8fcd-dfd193552704 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Received event network-vif-plugged-210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.459091] env[62952]: DEBUG oslo_concurrency.lockutils [req-b2ddaee9-3be3-42eb-83ba-e665a8c15394 req-6b55a99c-45c0-4788-8fcd-dfd193552704 service nova] Acquiring lock "9036f863-9d77-491b-8aa9-978d0c400708-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.459444] env[62952]: DEBUG oslo_concurrency.lockutils [req-b2ddaee9-3be3-42eb-83ba-e665a8c15394 req-6b55a99c-45c0-4788-8fcd-dfd193552704 service nova] Lock "9036f863-9d77-491b-8aa9-978d0c400708-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.460833] env[62952]: DEBUG oslo_concurrency.lockutils [req-b2ddaee9-3be3-42eb-83ba-e665a8c15394 req-6b55a99c-45c0-4788-8fcd-dfd193552704 service nova] Lock "9036f863-9d77-491b-8aa9-978d0c400708-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.460833] env[62952]: DEBUG nova.compute.manager [req-b2ddaee9-3be3-42eb-83ba-e665a8c15394 req-6b55a99c-45c0-4788-8fcd-dfd193552704 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] No waiting events found dispatching network-vif-plugged-210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1106.460833] env[62952]: WARNING nova.compute.manager [req-b2ddaee9-3be3-42eb-83ba-e665a8c15394 req-6b55a99c-45c0-4788-8fcd-dfd193552704 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Received unexpected event network-vif-plugged-210207ee-1211-4126-a2a8-7211dc7dfb11 for instance with vm_state building and task_state spawning. [ 1106.598729] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367656, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.679994] env[62952]: DEBUG nova.compute.manager [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Received event network-changed-2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.680329] env[62952]: DEBUG nova.compute.manager [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Refreshing instance network info cache due to event network-changed-2595a3f5-136a-4d28-83ff-9bcc6d05154e. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1106.680653] env[62952]: DEBUG oslo_concurrency.lockutils [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] Acquiring lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.680972] env[62952]: DEBUG oslo_concurrency.lockutils [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] Acquired lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.681461] env[62952]: DEBUG nova.network.neutron [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Refreshing network info cache for port 2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.756731] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553927} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.757018] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.757250] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1106.757839] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-36c6a1ee-d8f7-40a1-ac7d-511f04b53bfe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.767095] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1106.767095] env[62952]: value = "task-1367657" [ 1106.767095] env[62952]: _type = "Task" [ 1106.767095] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.775778] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367657, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.787121] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]526eef77-de94-e8c5-e439-bd761663e9fc, 'name': SearchDatastore_Task, 'duration_secs': 0.01228} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.787462] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.788147] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1106.788473] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-500aacf4-c532-4b5d-a588-d34b73ee72e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.795692] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1106.795692] env[62952]: value = "task-1367658" [ 1106.795692] env[62952]: _type = "Task" [ 1106.795692] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.804126] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.872765] env[62952]: DEBUG nova.network.neutron [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Updated VIF entry in instance network info cache for port 371d4d90-8873-4e1a-a839-449587dadfe0. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1106.873239] env[62952]: DEBUG nova.network.neutron [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Updating instance_info_cache with network_info: [{"id": "371d4d90-8873-4e1a-a839-449587dadfe0", "address": "fa:16:3e:3d:77:12", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap371d4d90-88", "ovs_interfaceid": "371d4d90-8873-4e1a-a839-449587dadfe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.924365] env[62952]: DEBUG nova.network.neutron [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Successfully updated port: 210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1106.930123] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "39eef6be-fe54-418b-b88d-104d082c3ca7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.930447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.930693] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "39eef6be-fe54-418b-b88d-104d082c3ca7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.930889] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.931091] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.933881] env[62952]: INFO nova.compute.manager [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Terminating instance [ 1106.938974] env[62952]: DEBUG nova.compute.manager [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1106.939295] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1106.941395] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a754f4f-5847-44c9-8715-377f9fa0a95c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.951456] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.952672] env[62952]: INFO nova.scheduler.client.report [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocation for migration c7775c81-116a-43e5-8781-e6537d043d8b [ 1106.953767] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3c06b1d-3ade-49cc-a832-17ed194bceb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.963159] env[62952]: DEBUG oslo_vmware.api [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1106.963159] env[62952]: value = "task-1367659" [ 1106.963159] env[62952]: _type = "Task" [ 1106.963159] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.972593] env[62952]: DEBUG oslo_vmware.api [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367659, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.097928] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367656, 'name': CreateVM_Task, 'duration_secs': 0.582726} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.098143] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1107.098870] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.099812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.099812] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1107.099812] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acc80af4-efe5-4fb5-90d4-e7c78ac02da6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.104526] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1107.104526] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52d60aca-9ba6-bde8-c8ec-aec95133bb9d" [ 1107.104526] env[62952]: _type = "Task" [ 1107.104526] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.112553] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d60aca-9ba6-bde8-c8ec-aec95133bb9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.277962] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367657, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124912} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.278279] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1107.279126] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0880aec-295a-4919-8d92-e57513bc297b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.301096] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.303769] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ed6b346-56c7-4c5c-87b7-6c249795fa7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.326895] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367658, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.328435] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1107.328435] env[62952]: value = "task-1367660" [ 1107.328435] env[62952]: _type = "Task" [ 1107.328435] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.337694] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367660, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.376234] env[62952]: DEBUG oslo_concurrency.lockutils [req-577ea5e2-af69-4f25-9207-58574d4a0529 req-f925a134-af6e-4bf6-9881-79c3cce88651 service nova] Releasing lock "refresh_cache-f302233b-2e1f-4c71-b000-0c9eb6530730" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.426796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "refresh_cache-9036f863-9d77-491b-8aa9-978d0c400708" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.426921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "refresh_cache-9036f863-9d77-491b-8aa9-978d0c400708" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.427093] env[62952]: DEBUG nova.network.neutron [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1107.443098] env[62952]: DEBUG nova.network.neutron [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updated VIF entry in instance network info cache for port 2595a3f5-136a-4d28-83ff-9bcc6d05154e. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1107.443515] env[62952]: DEBUG nova.network.neutron [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updating instance_info_cache with network_info: [{"id": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "address": "fa:16:3e:b0:9b:f1", "network": {"id": "bdb9dfb5-b6c2-4906-b90d-b3edd9d783dd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-701468107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8427597d4aae451e9375ade7c06b2bcf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2595a3f5-13", "ovs_interfaceid": "2595a3f5-136a-4d28-83ff-9bcc6d05154e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.459655] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.915s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.473491] env[62952]: DEBUG oslo_vmware.api [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367659, 'name': PowerOffVM_Task, 'duration_secs': 0.338432} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.474025] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1107.474209] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1107.474458] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5ab7475-0d60-465e-b177-edbb058afa6d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.614237] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d60aca-9ba6-bde8-c8ec-aec95133bb9d, 'name': SearchDatastore_Task, 'duration_secs': 0.055222} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.614558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.614769] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1107.615016] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.615181] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.615368] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1107.615624] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1fe1c928-3e6e-45df-9f99-572dadae57d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.623307] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1107.623474] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1107.624150] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6636b299-22e4-481b-945d-781ad1828d9f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.628892] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1107.628892] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5212d87c-b3a2-1a5a-2785-1c80e1375f7b" [ 1107.628892] env[62952]: _type = "Task" [ 1107.628892] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.636117] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5212d87c-b3a2-1a5a-2785-1c80e1375f7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.729748] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1107.729986] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1107.730197] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleting the datastore file [datastore1] 39eef6be-fe54-418b-b88d-104d082c3ca7 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1107.730506] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d9f5a9d-8600-4aaa-95e8-7b49ecd818c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.736840] env[62952]: DEBUG oslo_vmware.api [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1107.736840] env[62952]: value = "task-1367662" [ 1107.736840] env[62952]: _type = "Task" [ 1107.736840] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.747401] env[62952]: DEBUG oslo_vmware.api [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.776030] env[62952]: DEBUG nova.compute.manager [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Received event network-changed-210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.776340] env[62952]: DEBUG nova.compute.manager [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Refreshing instance network info cache due to event network-changed-210207ee-1211-4126-a2a8-7211dc7dfb11. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1107.776621] env[62952]: DEBUG oslo_concurrency.lockutils [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] Acquiring lock "refresh_cache-9036f863-9d77-491b-8aa9-978d0c400708" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.813424] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631529} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.813733] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1107.813981] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1107.814295] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d336423-1e14-4b31-9145-6610b2f2b951 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.820936] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1107.820936] env[62952]: value = "task-1367663" [ 1107.820936] env[62952]: _type = "Task" [ 1107.820936] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.829046] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.836829] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367660, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.946125] env[62952]: DEBUG oslo_concurrency.lockutils [req-f05fd1d5-8673-41fe-8353-7f6462f96b67 req-643ea7d5-9f66-4c60-99a7-a6ee8a536dbc service nova] Releasing lock "refresh_cache-287250a7-3354-49b7-9194-d88ae51afdaf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.959161] env[62952]: DEBUG nova.network.neutron [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1108.089666] env[62952]: DEBUG nova.network.neutron [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Updating instance_info_cache with network_info: [{"id": "210207ee-1211-4126-a2a8-7211dc7dfb11", "address": "fa:16:3e:ef:48:f6", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap210207ee-12", "ovs_interfaceid": "210207ee-1211-4126-a2a8-7211dc7dfb11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.139351] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5212d87c-b3a2-1a5a-2785-1c80e1375f7b, 'name': SearchDatastore_Task, 'duration_secs': 0.008924} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.140377] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bfe7cc2-a4d8-4af3-af0d-842efed2b64d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.145294] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1108.145294] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52670acd-b072-8210-aaeb-e4dc17c5a54d" [ 1108.145294] env[62952]: _type = "Task" [ 1108.145294] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.152469] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52670acd-b072-8210-aaeb-e4dc17c5a54d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.246057] env[62952]: DEBUG oslo_vmware.api [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367662, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159702} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.246335] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1108.246540] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1108.246733] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1108.246908] env[62952]: INFO nova.compute.manager [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Took 1.31 seconds to destroy the instance on the hypervisor. [ 1108.247170] env[62952]: DEBUG oslo.service.loopingcall [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.247369] env[62952]: DEBUG nova.compute.manager [-] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1108.247463] env[62952]: DEBUG nova.network.neutron [-] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1108.331008] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065729} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.333472] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1108.334421] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd124d11-0b0d-42dc-9986-ded09ab912a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.341490] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367660, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.358951] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.359231] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-563210d1-00a8-4e9f-8430-6fad33dc404a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.378801] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1108.378801] env[62952]: value = "task-1367664" [ 1108.378801] env[62952]: _type = "Task" [ 1108.378801] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.386757] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367664, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.570494] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.570795] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.571042] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.571242] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.571423] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.573949] env[62952]: INFO nova.compute.manager [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Terminating instance [ 1108.576345] env[62952]: DEBUG nova.compute.manager [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1108.576546] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.577386] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d874a20-c209-4562-a2cb-d336bb8c6609 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.585584] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.585886] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6757eb9-3ca2-423c-aa46-e2c0c2f7203e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.592082] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1108.592082] env[62952]: value = "task-1367665" [ 1108.592082] env[62952]: _type = "Task" [ 1108.592082] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.592593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "refresh_cache-9036f863-9d77-491b-8aa9-978d0c400708" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.592886] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Instance network_info: |[{"id": "210207ee-1211-4126-a2a8-7211dc7dfb11", "address": "fa:16:3e:ef:48:f6", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap210207ee-12", "ovs_interfaceid": "210207ee-1211-4126-a2a8-7211dc7dfb11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1108.595932] env[62952]: DEBUG oslo_concurrency.lockutils [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] Acquired lock "refresh_cache-9036f863-9d77-491b-8aa9-978d0c400708" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.596132] env[62952]: DEBUG nova.network.neutron [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Refreshing network info cache for port 210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1108.597293] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:48:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '210207ee-1211-4126-a2a8-7211dc7dfb11', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.605881] env[62952]: DEBUG oslo.service.loopingcall [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.606443] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1108.607144] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9d912f0-e48e-41b9-902e-bd5190520aac {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.626124] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.631157] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.631157] env[62952]: value = "task-1367666" [ 1108.631157] env[62952]: _type = "Task" [ 1108.631157] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.638777] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367666, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.656390] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52670acd-b072-8210-aaeb-e4dc17c5a54d, 'name': SearchDatastore_Task, 'duration_secs': 0.018182} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.656654] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.656911] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] f302233b-2e1f-4c71-b000-0c9eb6530730/f302233b-2e1f-4c71-b000-0c9eb6530730.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1108.657180] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6554d8f8-5dc4-4a57-970f-a46ccc4fe843 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.664500] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1108.664500] env[62952]: value = "task-1367667" [ 1108.664500] env[62952]: _type = "Task" [ 1108.664500] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.672717] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.844536] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367660, 'name': ReconfigVM_Task, 'duration_secs': 1.081598} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.845085] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to attach disk [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.846089] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-050fbaac-c43f-481f-8738-efc0d79e7dc1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.855372] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1108.855372] env[62952]: value = "task-1367668" [ 1108.855372] env[62952]: _type = "Task" [ 1108.855372] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.867524] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367668, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.888618] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367664, 'name': ReconfigVM_Task, 'duration_secs': 0.288138} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.888998] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfigured VM instance instance-0000005c to attach disk [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d/f5e5c5c1-e2af-40e0-a957-67b0fd22718d.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.890528] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encrypted': False, 'device_name': '/dev/sda', 'boot_index': 0, 'encryption_secret_uuid': None, 'encryption_options': None, 'guest_format': None, 'device_type': 'disk', 'encryption_format': None, 'size': 0, 'disk_bus': None, 'image_id': 'e04fcbd0-b3d8-461a-890d-eda04a2a1e77'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'attachment_id': '490e9c46-0bdc-458e-8d4b-79f0808d2c73', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'}, 'guest_format': None, 'device_type': None, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62952) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1108.890748] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1108.890950] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1108.891762] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2083790-0298-4ee3-8928-6a09affb42f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.907375] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b10d53-a6d6-4f5b-ae68-e32d52a973bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.932347] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a/volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.932691] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32228bf8-0441-4011-ac03-d2e1b804dbde {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.954327] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1108.954327] env[62952]: value = "task-1367669" [ 1108.954327] env[62952]: _type = "Task" [ 1108.954327] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.966058] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.082937] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.083601] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.083601] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1109.104689] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367665, 'name': PowerOffVM_Task, 'duration_secs': 0.192275} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.105064] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.105277] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.105552] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7ed4eac-7866-429e-9ea4-73b5bc89a247 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.143177] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367666, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.176782] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367667, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.233025] env[62952]: DEBUG nova.network.neutron [-] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.237902] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.238101] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.238288] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleting the datastore file [datastore1] 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.238559] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fdf166c-2638-474f-8be1-a4bce4dffd46 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.249260] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for the task: (returnval){ [ 1109.249260] env[62952]: value = "task-1367671" [ 1109.249260] env[62952]: _type = "Task" [ 1109.249260] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.258865] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367671, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.368476] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367668, 'name': Rename_Task, 'duration_secs': 0.237657} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.370793] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1109.371068] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6460ad2-e861-439e-8b55-fcbe2356a4f2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.376991] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1109.376991] env[62952]: value = "task-1367672" [ 1109.376991] env[62952]: _type = "Task" [ 1109.376991] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.384522] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.405536] env[62952]: DEBUG nova.network.neutron [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Updated VIF entry in instance network info cache for port 210207ee-1211-4126-a2a8-7211dc7dfb11. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1109.405907] env[62952]: DEBUG nova.network.neutron [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Updating instance_info_cache with network_info: [{"id": "210207ee-1211-4126-a2a8-7211dc7dfb11", "address": "fa:16:3e:ef:48:f6", "network": {"id": "607b04d0-9352-4465-ba15-8e704d98ee5f", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-221041815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9657a113032417fb4b97716a772f8ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap210207ee-12", "ovs_interfaceid": "210207ee-1211-4126-a2a8-7211dc7dfb11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.464148] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367669, 'name': ReconfigVM_Task, 'duration_secs': 0.404835} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.464501] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a/volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1109.469288] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed5162fe-10bb-4b44-89ce-3cf69b86c857 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.484084] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1109.484084] env[62952]: value = "task-1367673" [ 1109.484084] env[62952]: _type = "Task" [ 1109.484084] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.491967] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367673, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.642197] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367666, 'name': CreateVM_Task, 'duration_secs': 0.606516} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.642552] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.643175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.643405] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.643741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.644024] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c38230c-e1b7-4d11-9c31-3c903b02ae43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.648274] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1109.648274] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52ee1e92-232d-37bb-3b5e-df3c52be0016" [ 1109.648274] env[62952]: _type = "Task" [ 1109.648274] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.656178] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ee1e92-232d-37bb-3b5e-df3c52be0016, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.673435] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367667, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.665288} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.673687] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] f302233b-2e1f-4c71-b000-0c9eb6530730/f302233b-2e1f-4c71-b000-0c9eb6530730.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1109.673900] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1109.674195] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd151621-e3cc-4876-b2af-555e370a70fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.680240] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1109.680240] env[62952]: value = "task-1367674" [ 1109.680240] env[62952]: _type = "Task" [ 1109.680240] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.687644] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.734915] env[62952]: INFO nova.compute.manager [-] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Took 1.49 seconds to deallocate network for instance. [ 1109.755358] env[62952]: DEBUG oslo_vmware.api [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Task: {'id': task-1367671, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229764} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.755612] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.755802] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.755987] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.756188] env[62952]: INFO nova.compute.manager [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1109.756439] env[62952]: DEBUG oslo.service.loopingcall [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.756632] env[62952]: DEBUG nova.compute.manager [-] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.756728] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1109.801882] env[62952]: DEBUG nova.compute.manager [req-be7d0885-4632-429e-a439-47ccb92d199d req-b1287a17-31c0-4e7b-9bcb-0660b8745f99 service nova] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Received event network-vif-deleted-a6a41344-0103-472c-a33e-a85dc9674a09 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.889100] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367672, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.908406] env[62952]: DEBUG oslo_concurrency.lockutils [req-1673f2d0-8bca-4250-b431-cdbb4012d839 req-b5e8ee24-bf1c-4af7-8b5f-af061dc86047 service nova] Releasing lock "refresh_cache-9036f863-9d77-491b-8aa9-978d0c400708" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.993705] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367673, 'name': ReconfigVM_Task, 'duration_secs': 0.144631} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.993966] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1109.994593] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-412ec86e-af30-472b-9285-4968e50eb661 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.001495] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1110.001495] env[62952]: value = "task-1367675" [ 1110.001495] env[62952]: _type = "Task" [ 1110.001495] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.011181] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367675, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.158689] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52ee1e92-232d-37bb-3b5e-df3c52be0016, 'name': SearchDatastore_Task, 'duration_secs': 0.009551} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.158987] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.159243] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.159484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.159637] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.159816] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.160318] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eeb3b2bf-d4e5-4b02-9736-2eba718aa2d0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.167414] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.167598] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1110.168290] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f134b01-b618-482c-a2ef-25b1cbf17200 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.173021] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1110.173021] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5227807b-3ff3-d721-55bb-82469e61449e" [ 1110.173021] env[62952]: _type = "Task" [ 1110.173021] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.179915] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5227807b-3ff3-d721-55bb-82469e61449e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.187969] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253778} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.188234] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1110.188972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9b8337-5c00-47d7-80c0-6668da887bc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.210525] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] f302233b-2e1f-4c71-b000-0c9eb6530730/f302233b-2e1f-4c71-b000-0c9eb6530730.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1110.210789] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-484ac8fc-0746-4847-9669-c5149e584d07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.231532] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1110.231532] env[62952]: value = "task-1367676" [ 1110.231532] env[62952]: _type = "Task" [ 1110.231532] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.239594] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.240530] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.240807] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.241052] env[62952]: DEBUG nova.objects.instance [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'resources' on Instance uuid 39eef6be-fe54-418b-b88d-104d082c3ca7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.387923] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367672, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.472733] env[62952]: DEBUG nova.network.neutron [-] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.511843] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367675, 'name': Rename_Task, 'duration_secs': 0.435966} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.512180] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1110.512470] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b22de67d-f710-40b5-adb2-4873add2570b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.519070] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1110.519070] env[62952]: value = "task-1367677" [ 1110.519070] env[62952]: _type = "Task" [ 1110.519070] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.526543] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.592023] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1110.592356] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.592543] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.592701] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.592857] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.593022] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.593182] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.593316] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1110.593476] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.683273] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5227807b-3ff3-d721-55bb-82469e61449e, 'name': SearchDatastore_Task, 'duration_secs': 0.008241} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.684100] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a1f1a4a-bafd-4978-b706-89c2ce883130 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.688852] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1110.688852] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]527a57b4-1a31-c058-148a-698f80f5aff2" [ 1110.688852] env[62952]: _type = "Task" [ 1110.688852] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.696175] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527a57b4-1a31-c058-148a-698f80f5aff2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.745826] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367676, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.871827] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd68a175-1242-4b1e-aaaa-356051921ddc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.882252] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4f3410-08ae-4f2c-a79f-ca57c3a7bc62 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.889545] env[62952]: DEBUG oslo_vmware.api [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367672, 'name': PowerOnVM_Task, 'duration_secs': 1.162261} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.913693] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1110.913950] env[62952]: INFO nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Took 9.28 seconds to spawn the instance on the hypervisor. [ 1110.914173] env[62952]: DEBUG nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1110.915273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47b5e08-1b4c-4f99-bed9-b453e195a80c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.918619] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0bf067-7098-4560-aff3-3e2b87f8eed5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.930210] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ca74ae-83b1-44f5-a731-a59924c7e8d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.943287] env[62952]: DEBUG nova.compute.provider_tree [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.976160] env[62952]: INFO nova.compute.manager [-] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Took 1.22 seconds to deallocate network for instance. [ 1111.028694] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367677, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.096440] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.201203] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]527a57b4-1a31-c058-148a-698f80f5aff2, 'name': SearchDatastore_Task, 'duration_secs': 0.314065} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.201459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.201720] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 9036f863-9d77-491b-8aa9-978d0c400708/9036f863-9d77-491b-8aa9-978d0c400708.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1111.202026] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a861acf-3e9a-49a3-9a85-14b7aa5b0eb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.208180] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1111.208180] env[62952]: value = "task-1367678" [ 1111.208180] env[62952]: _type = "Task" [ 1111.208180] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.215555] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.241141] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367676, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.437034] env[62952]: INFO nova.compute.manager [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Took 16.16 seconds to build instance. [ 1111.446666] env[62952]: DEBUG nova.scheduler.client.report [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1111.483506] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.531406] env[62952]: DEBUG oslo_vmware.api [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367677, 'name': PowerOnVM_Task, 'duration_secs': 0.553927} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.531406] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1111.531406] env[62952]: DEBUG nova.compute.manager [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1111.531406] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038eb276-ee28-466e-a04a-ab530fb9085b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.717298] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367678, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.742594] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367676, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.830745] env[62952]: DEBUG nova.compute.manager [req-463aff03-852f-4e7c-bec5-41ce9f6e5768 req-961c1b72-af27-4fd7-a264-5bedfbc34993 service nova] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Received event network-vif-deleted-69251416-59fd-45c5-a937-fe98301eb5a7 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.939233] env[62952]: DEBUG oslo_concurrency.lockutils [None req-7f423baf-c4fb-404b-829e-5837962119db tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.671s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.951205] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.955781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.859s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.955781] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.955781] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1111.958661] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.473s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.958661] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.960026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1823483-236c-4950-b018-633fe96fe241 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.970772] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1954c788-16c3-4396-a0dc-eab0a4340c86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.987403] env[62952]: INFO nova.scheduler.client.report [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Deleted allocations for instance 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed [ 1111.991845] env[62952]: INFO nova.scheduler.client.report [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted allocations for instance 39eef6be-fe54-418b-b88d-104d082c3ca7 [ 1111.993702] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eeb75f2-14c8-4536-a33f-93444715b093 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.013645] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe831a04-1946-4a18-b4f3-dd4d4ca1817d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.048770] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179988MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1112.048933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.049274] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.055749] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.103268] env[62952]: DEBUG nova.compute.manager [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Received event network-changed-612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.103479] env[62952]: DEBUG nova.compute.manager [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Refreshing instance network info cache due to event network-changed-612c5fea-4183-448f-9326-6ed1888b591d. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1112.103703] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.103854] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.104058] env[62952]: DEBUG nova.network.neutron [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Refreshing network info cache for port 612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.218571] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367678, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.243089] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367676, 'name': ReconfigVM_Task, 'duration_secs': 1.563195} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.243720] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Reconfigured VM instance instance-00000066 to attach disk [datastore1] f302233b-2e1f-4c71-b000-0c9eb6530730/f302233b-2e1f-4c71-b000-0c9eb6530730.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.244386] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5044b43c-eb56-437b-995d-6167dcb80898 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.254830] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1112.254830] env[62952]: value = "task-1367679" [ 1112.254830] env[62952]: _type = "Task" [ 1112.254830] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.264522] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367679, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.512372] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d580ea7a-cb88-4bbd-a055-1d75adf28a56 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "39eef6be-fe54-418b-b88d-104d082c3ca7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.582s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.516077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-91bbfeb0-9145-427e-845e-b058f3899a71 tempest-DeleteServersTestJSON-687693937 tempest-DeleteServersTestJSON-687693937-project-member] Lock "9c18c8c7-bc97-4a84-a86d-e23f8cce20ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.945s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.718967] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367678, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.110393} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.719274] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 9036f863-9d77-491b-8aa9-978d0c400708/9036f863-9d77-491b-8aa9-978d0c400708.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1112.719494] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1112.719758] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e38b131e-a6a7-4d6c-ad25-d7fcbd1d1562 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.725468] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1112.725468] env[62952]: value = "task-1367680" [ 1112.725468] env[62952]: _type = "Task" [ 1112.725468] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.733256] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.763854] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367679, 'name': Rename_Task, 'duration_secs': 0.445884} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.764524] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.765203] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e5dbeaa-0b96-4bca-979a-622514138c71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.771750] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1112.771750] env[62952]: value = "task-1367681" [ 1112.771750] env[62952]: _type = "Task" [ 1112.771750] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.782428] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.866807] env[62952]: DEBUG nova.network.neutron [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updated VIF entry in instance network info cache for port 612c5fea-4183-448f-9326-6ed1888b591d. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1112.867190] env[62952]: DEBUG nova.network.neutron [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.077710] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.077873] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 532a2bcd-5d6a-4fa9-abc9-a048cc915fda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078037] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 75d640ed-c41a-4761-8867-191d8b3e1f79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078152] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance f5e5c5c1-e2af-40e0-a957-67b0fd22718d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078267] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 287250a7-3354-49b7-9194-d88ae51afdaf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078384] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ff8301b2-c5cb-4e46-8f5c-98577dfd4835 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078488] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance f302233b-2e1f-4c71-b000-0c9eb6530730 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078596] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 9036f863-9d77-491b-8aa9-978d0c400708 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1113.078786] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1113.078922] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1113.216010] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da166d76-eb63-47ee-a367-1921bc67af83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.223471] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4ee910-9bab-4c6c-a6d2-9ee3f0ccad43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.237675] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.263707] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db81cde3-4f1a-470a-b003-3184dc3d7f73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.271333] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12247d4c-d90b-4770-8e1a-8a731d3a558e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.289614] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.294512] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.369981] env[62952]: DEBUG oslo_concurrency.lockutils [req-9a0c5b63-ba1b-42b5-a607-75bc2504f574 req-60a34a3d-b911-44fd-b8dc-d955da4dcb77 service nova] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.604479] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.604802] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.737116] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.786994] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367681, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.796095] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.094787] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.095302] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.107881] env[62952]: INFO nova.compute.manager [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Detaching volume 3aaf2335-5c53-4729-abe8-cf69d9aad29a [ 1114.156228] env[62952]: INFO nova.virt.block_device [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Attempting to driver detach volume 3aaf2335-5c53-4729-abe8-cf69d9aad29a from mountpoint /dev/sdb [ 1114.156848] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1114.156848] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1114.157699] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb54c20-170a-4a25-83a9-0d393a08a183 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.182584] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b314a1-bf1f-4add-8b54-b6ac1e7ecc03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.190965] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27afa29e-54c6-48fc-8649-1aae304d3424 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.212773] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de347541-09b8-4f93-8fdd-463ab3de0808 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.234265] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] The volume has not been displaced from its original location: [datastore2] volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a/volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1114.243527] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.249878] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8c2d7ed-08ee-4eca-8c08-2fa83a5d7a0e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.278335] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.414333} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.279557] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1114.285239] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2496d9da-0c6b-4195-a728-5998bd0893ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.290464] env[62952]: DEBUG oslo_vmware.api [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1114.290464] env[62952]: value = "task-1367682" [ 1114.290464] env[62952]: _type = "Task" [ 1114.290464] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.315665] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1114.315994] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.267s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.316397] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367681, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.331370] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 9036f863-9d77-491b-8aa9-978d0c400708/9036f863-9d77-491b-8aa9-978d0c400708.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.332538] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.277s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.332867] env[62952]: DEBUG nova.objects.instance [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62952) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1114.336683] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae1b9c08-9eaf-4540-8eea-4a46223d059a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.361905] env[62952]: DEBUG oslo_vmware.api [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.368476] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1114.368476] env[62952]: value = "task-1367683" [ 1114.368476] env[62952]: _type = "Task" [ 1114.368476] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.377061] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367683, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.599869] env[62952]: INFO nova.compute.manager [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Detaching volume ac6ea8d6-f93e-490a-9d4e-c3c6816c1351 [ 1114.634739] env[62952]: INFO nova.virt.block_device [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Attempting to driver detach volume ac6ea8d6-f93e-490a-9d4e-c3c6816c1351 from mountpoint /dev/sdb [ 1114.634961] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1114.635177] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290968', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'name': 'volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '75d640ed-c41a-4761-8867-191d8b3e1f79', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'serial': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1114.636091] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b54c00-6a48-4783-aa17-ed9b6a729797 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.657569] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c067e5-0a96-4514-bf78-b58abc0f4dd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.664363] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40213d7b-6bd3-4704-8d7b-372afe1a2f6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.684165] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bd20b2-5510-4fc7-8034-600a51384539 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.700882] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] The volume has not been displaced from its original location: [datastore2] volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351/volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1114.706088] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.706419] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48b2a253-ff14-43ed-8fdd-fb92d60516fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.724113] env[62952]: DEBUG oslo_vmware.api [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1114.724113] env[62952]: value = "task-1367684" [ 1114.724113] env[62952]: _type = "Task" [ 1114.724113] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.732065] env[62952]: DEBUG oslo_vmware.api [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.787781] env[62952]: DEBUG oslo_vmware.api [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367681, 'name': PowerOnVM_Task, 'duration_secs': 1.530275} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.788107] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1114.788357] env[62952]: INFO nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Took 10.88 seconds to spawn the instance on the hypervisor. [ 1114.788592] env[62952]: DEBUG nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.789387] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bfd7bd-0323-4ca2-9501-f48cd61eb30b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.803841] env[62952]: DEBUG oslo_vmware.api [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367682, 'name': ReconfigVM_Task, 'duration_secs': 0.374534} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.804452] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1114.809303] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02127c31-d6db-4f44-bfd7-5634cd4e66b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.824144] env[62952]: DEBUG oslo_vmware.api [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1114.824144] env[62952]: value = "task-1367685" [ 1114.824144] env[62952]: _type = "Task" [ 1114.824144] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.832878] env[62952]: DEBUG oslo_vmware.api [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.878949] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367683, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.234862] env[62952]: DEBUG oslo_vmware.api [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367684, 'name': ReconfigVM_Task, 'duration_secs': 0.295049} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.235239] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1115.240091] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3876889-f1e1-49c3-9245-d3915f4f14eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.254958] env[62952]: DEBUG oslo_vmware.api [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1115.254958] env[62952]: value = "task-1367686" [ 1115.254958] env[62952]: _type = "Task" [ 1115.254958] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.263423] env[62952]: DEBUG oslo_vmware.api [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.312133] env[62952]: INFO nova.compute.manager [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Took 19.72 seconds to build instance. [ 1115.335546] env[62952]: DEBUG oslo_vmware.api [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367685, 'name': ReconfigVM_Task, 'duration_secs': 0.176277} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.335871] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290978', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'name': 'volume-3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f5e5c5c1-e2af-40e0-a957-67b0fd22718d', 'attached_at': '', 'detached_at': '', 'volume_id': '3aaf2335-5c53-4729-abe8-cf69d9aad29a', 'serial': '3aaf2335-5c53-4729-abe8-cf69d9aad29a'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.365195] env[62952]: DEBUG oslo_concurrency.lockutils [None req-30995a67-89c7-487d-a109-8c31eca03862 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.033s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.379467] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367683, 'name': ReconfigVM_Task, 'duration_secs': 0.589989} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.380315] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 9036f863-9d77-491b-8aa9-978d0c400708/9036f863-9d77-491b-8aa9-978d0c400708.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.380949] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e7167f6-da7a-4ef8-b4b0-6f1949ee0d2b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.387810] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1115.387810] env[62952]: value = "task-1367687" [ 1115.387810] env[62952]: _type = "Task" [ 1115.387810] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.396020] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367687, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.764318] env[62952]: DEBUG oslo_vmware.api [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367686, 'name': ReconfigVM_Task, 'duration_secs': 0.225293} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.764630] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290968', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'name': 'volume-ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '75d640ed-c41a-4761-8867-191d8b3e1f79', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351', 'serial': 'ac6ea8d6-f93e-490a-9d4e-c3c6816c1351'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.814738] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ef3f4f10-95df-48d5-84d9-23497d7bcc81 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.229s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.880820] env[62952]: DEBUG nova.objects.instance [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'flavor' on Instance uuid f5e5c5c1-e2af-40e0-a957-67b0fd22718d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.900493] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367687, 'name': Rename_Task, 'duration_secs': 0.156745} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.900772] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1115.901034] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-884eb2b9-b232-489d-a09f-0df240806f66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.907242] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1115.907242] env[62952]: value = "task-1367688" [ 1115.907242] env[62952]: _type = "Task" [ 1115.907242] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.914797] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.132267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "f302233b-2e1f-4c71-b000-0c9eb6530730" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.132513] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.132769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "f302233b-2e1f-4c71-b000-0c9eb6530730-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.132968] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.133161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.135368] env[62952]: INFO nova.compute.manager [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Terminating instance [ 1116.137140] env[62952]: DEBUG nova.compute.manager [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1116.137387] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1116.138226] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4ae7d6-7563-49aa-9fd1-2416a3d789f6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.146535] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1116.146781] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b376017-d4fb-45e3-8631-3a1c05cf55b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.152676] env[62952]: DEBUG oslo_vmware.api [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1116.152676] env[62952]: value = "task-1367689" [ 1116.152676] env[62952]: _type = "Task" [ 1116.152676] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.160629] env[62952]: DEBUG oslo_vmware.api [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.306961] env[62952]: DEBUG nova.objects.instance [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'flavor' on Instance uuid 75d640ed-c41a-4761-8867-191d8b3e1f79 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.416607] env[62952]: DEBUG oslo_vmware.api [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367688, 'name': PowerOnVM_Task, 'duration_secs': 0.486321} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.416909] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1116.417126] env[62952]: INFO nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Took 10.25 seconds to spawn the instance on the hypervisor. [ 1116.417312] env[62952]: DEBUG nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1116.418097] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bcc2cb-cd48-4443-abb9-77bbc6621047 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.662829] env[62952]: DEBUG oslo_vmware.api [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367689, 'name': PowerOffVM_Task, 'duration_secs': 0.20989} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.663145] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1116.663325] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1116.663571] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff523bad-3ae8-44a3-91d1-4c12b563fde7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.747834] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1116.747834] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1116.747834] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore1] f302233b-2e1f-4c71-b000-0c9eb6530730 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.748075] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6272e33e-712e-45cf-8a0e-cba0c8d31714 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.754929] env[62952]: DEBUG oslo_vmware.api [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1116.754929] env[62952]: value = "task-1367691" [ 1116.754929] env[62952]: _type = "Task" [ 1116.754929] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.762359] env[62952]: DEBUG oslo_vmware.api [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367691, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.889227] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a602d5c-970d-4fce-88af-581beee1a7b7 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.284s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.933692] env[62952]: INFO nova.compute.manager [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Took 16.86 seconds to build instance. [ 1117.265029] env[62952]: DEBUG oslo_vmware.api [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367691, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15924} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.265218] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.265420] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1117.265608] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1117.265830] env[62952]: INFO nova.compute.manager [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1117.266129] env[62952]: DEBUG oslo.service.loopingcall [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.266374] env[62952]: DEBUG nova.compute.manager [-] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1117.266481] env[62952]: DEBUG nova.network.neutron [-] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1117.314461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d4e89402-1d3f-41cc-a4ec-58419322d2b0 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.219s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.435800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a343f79f-fd3e-4512-998a-1ba24b2d89ac tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "9036f863-9d77-491b-8aa9-978d0c400708" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.364s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.605877] env[62952]: DEBUG nova.compute.manager [req-a9d0fa22-f2ab-4edb-895f-05497387917d req-8d66c672-5843-402d-84ae-f8f21abc23c0 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Received event network-vif-deleted-371d4d90-8873-4e1a-a839-449587dadfe0 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1117.606125] env[62952]: INFO nova.compute.manager [req-a9d0fa22-f2ab-4edb-895f-05497387917d req-8d66c672-5843-402d-84ae-f8f21abc23c0 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Neutron deleted interface 371d4d90-8873-4e1a-a839-449587dadfe0; detaching it from the instance and deleting it from the info cache [ 1117.606305] env[62952]: DEBUG nova.network.neutron [req-a9d0fa22-f2ab-4edb-895f-05497387917d req-8d66c672-5843-402d-84ae-f8f21abc23c0 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.722700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "9036f863-9d77-491b-8aa9-978d0c400708" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.723012] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "9036f863-9d77-491b-8aa9-978d0c400708" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.723209] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "9036f863-9d77-491b-8aa9-978d0c400708-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.723402] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "9036f863-9d77-491b-8aa9-978d0c400708-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.723577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "9036f863-9d77-491b-8aa9-978d0c400708-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.728533] env[62952]: INFO nova.compute.manager [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Terminating instance [ 1117.730522] env[62952]: DEBUG nova.compute.manager [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.730861] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.731775] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa53e87b-3f42-42ed-a697-bf14b838a1ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.739232] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.739414] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89862eaf-d98a-41b1-b088-c4c136ff82ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.749440] env[62952]: DEBUG oslo_vmware.api [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1117.749440] env[62952]: value = "task-1367692" [ 1117.749440] env[62952]: _type = "Task" [ 1117.749440] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.757096] env[62952]: DEBUG oslo_vmware.api [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.952351] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.952856] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.952955] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.953128] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.953317] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.955407] env[62952]: INFO nova.compute.manager [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Terminating instance [ 1117.957110] env[62952]: DEBUG nova.compute.manager [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.957310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.958132] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00423919-8320-402d-903d-2a30b7896955 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.965862] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.966120] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5b02ff1-3866-44af-bfbd-7c5fed03c171 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.971706] env[62952]: DEBUG oslo_vmware.api [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1117.971706] env[62952]: value = "task-1367693" [ 1117.971706] env[62952]: _type = "Task" [ 1117.971706] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.979317] env[62952]: DEBUG oslo_vmware.api [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.051292] env[62952]: DEBUG nova.network.neutron [-] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.110052] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddc2ef81-533c-4644-b28d-5a8b6ccd529a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.119546] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe133e27-b569-4b96-bdfb-3d19fdb46c64 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.147573] env[62952]: DEBUG nova.compute.manager [req-a9d0fa22-f2ab-4edb-895f-05497387917d req-8d66c672-5843-402d-84ae-f8f21abc23c0 service nova] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Detach interface failed, port_id=371d4d90-8873-4e1a-a839-449587dadfe0, reason: Instance f302233b-2e1f-4c71-b000-0c9eb6530730 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1118.259422] env[62952]: DEBUG oslo_vmware.api [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367692, 'name': PowerOffVM_Task, 'duration_secs': 0.207149} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.261197] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.261197] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.261197] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aed71220-9790-4241-9be1-9dfd7480b914 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.320151] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.320375] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.320580] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleting the datastore file [datastore1] 9036f863-9d77-491b-8aa9-978d0c400708 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.320843] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25f5239c-5304-4ddf-bca9-82c14b097ae8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.326995] env[62952]: DEBUG oslo_vmware.api [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for the task: (returnval){ [ 1118.326995] env[62952]: value = "task-1367695" [ 1118.326995] env[62952]: _type = "Task" [ 1118.326995] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.334220] env[62952]: DEBUG oslo_vmware.api [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.346700] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.346936] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.347158] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "75d640ed-c41a-4761-8867-191d8b3e1f79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.347352] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.347525] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.349696] env[62952]: INFO nova.compute.manager [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Terminating instance [ 1118.351426] env[62952]: DEBUG nova.compute.manager [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1118.351643] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1118.352378] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1886e99d-85a1-4466-bb2a-757ca11a3d47 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.358854] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1118.359083] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f74494c-965a-4927-9043-367eae6b422c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.364498] env[62952]: DEBUG oslo_vmware.api [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1118.364498] env[62952]: value = "task-1367696" [ 1118.364498] env[62952]: _type = "Task" [ 1118.364498] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.372772] env[62952]: DEBUG oslo_vmware.api [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367696, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.484344] env[62952]: DEBUG oslo_vmware.api [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367693, 'name': PowerOffVM_Task, 'duration_secs': 0.184132} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.484611] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.484910] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.485264] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abb1a16c-5528-49eb-a9c1-aa1e22348d61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.554458] env[62952]: INFO nova.compute.manager [-] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Took 1.29 seconds to deallocate network for instance. [ 1118.558430] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.558697] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.558889] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore1] f5e5c5c1-e2af-40e0-a957-67b0fd22718d {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.561705] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b69d1d85-1869-40a6-a8fe-ff3c261f952d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.569206] env[62952]: DEBUG oslo_vmware.api [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1118.569206] env[62952]: value = "task-1367698" [ 1118.569206] env[62952]: _type = "Task" [ 1118.569206] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.577581] env[62952]: DEBUG oslo_vmware.api [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.836537] env[62952]: DEBUG oslo_vmware.api [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Task: {'id': task-1367695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174829} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.836902] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1118.837119] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1118.837310] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1118.837492] env[62952]: INFO nova.compute.manager [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1118.837734] env[62952]: DEBUG oslo.service.loopingcall [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1118.837927] env[62952]: DEBUG nova.compute.manager [-] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1118.838063] env[62952]: DEBUG nova.network.neutron [-] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1118.877401] env[62952]: DEBUG oslo_vmware.api [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367696, 'name': PowerOffVM_Task, 'duration_secs': 0.158096} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.877770] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.877967] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.878242] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57d7ab8e-4220-4eb8-b42c-8f22354faeb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.941502] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.941853] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.942074] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleting the datastore file [datastore2] 75d640ed-c41a-4761-8867-191d8b3e1f79 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.942357] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b96b8440-b3d7-4b65-8413-d35139b217d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.949187] env[62952]: DEBUG oslo_vmware.api [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1118.949187] env[62952]: value = "task-1367700" [ 1118.949187] env[62952]: _type = "Task" [ 1118.949187] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.957413] env[62952]: DEBUG oslo_vmware.api [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.065504] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.065863] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.065978] env[62952]: DEBUG nova.objects.instance [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid f302233b-2e1f-4c71-b000-0c9eb6530730 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.082119] env[62952]: DEBUG oslo_vmware.api [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17833} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.082561] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.082648] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.082794] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.082972] env[62952]: INFO nova.compute.manager [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1119.083232] env[62952]: DEBUG oslo.service.loopingcall [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.083566] env[62952]: DEBUG nova.compute.manager [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.083566] env[62952]: DEBUG nova.network.neutron [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1119.459764] env[62952]: DEBUG oslo_vmware.api [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139888} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.460038] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.460236] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.460424] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.461120] env[62952]: INFO nova.compute.manager [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1119.461120] env[62952]: DEBUG oslo.service.loopingcall [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.461120] env[62952]: DEBUG nova.compute.manager [-] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.461120] env[62952]: DEBUG nova.network.neutron [-] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1119.561892] env[62952]: DEBUG nova.network.neutron [-] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.650641] env[62952]: DEBUG nova.compute.manager [req-cdca6912-515b-4744-b600-44a7ece9d4ab req-1e60b393-856a-46ed-97c6-c4712dabef27 service nova] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Received event network-vif-deleted-210207ee-1211-4126-a2a8-7211dc7dfb11 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.650870] env[62952]: DEBUG nova.compute.manager [req-cdca6912-515b-4744-b600-44a7ece9d4ab req-1e60b393-856a-46ed-97c6-c4712dabef27 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Received event network-vif-deleted-be83803d-9485-4459-882a-5dc9383e0f62 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.651048] env[62952]: INFO nova.compute.manager [req-cdca6912-515b-4744-b600-44a7ece9d4ab req-1e60b393-856a-46ed-97c6-c4712dabef27 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Neutron deleted interface be83803d-9485-4459-882a-5dc9383e0f62; detaching it from the instance and deleting it from the info cache [ 1119.651223] env[62952]: DEBUG nova.network.neutron [req-cdca6912-515b-4744-b600-44a7ece9d4ab req-1e60b393-856a-46ed-97c6-c4712dabef27 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.726906] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64606533-7ad7-4567-8632-9b40dc6c22af {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.735066] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e6b9db-1014-45ad-83bc-d99dec9468de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.772454] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fe24a5-a6d5-4804-a1ae-a05ee24dbd39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.780084] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2b384a-e3b3-4a42-ba28-1dda45a06e1b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.795234] env[62952]: DEBUG nova.compute.provider_tree [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.962765] env[62952]: DEBUG nova.compute.manager [req-a1181df2-6023-4da3-958c-b2ed23013fd6 req-063d5c1d-0806-4992-9f82-037c5423a9c6 service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Received event network-vif-deleted-6e4543b9-4601-4109-b4fc-14eee838c035 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.963021] env[62952]: INFO nova.compute.manager [req-a1181df2-6023-4da3-958c-b2ed23013fd6 req-063d5c1d-0806-4992-9f82-037c5423a9c6 service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Neutron deleted interface 6e4543b9-4601-4109-b4fc-14eee838c035; detaching it from the instance and deleting it from the info cache [ 1119.963197] env[62952]: DEBUG nova.network.neutron [req-a1181df2-6023-4da3-958c-b2ed23013fd6 req-063d5c1d-0806-4992-9f82-037c5423a9c6 service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.066164] env[62952]: INFO nova.compute.manager [-] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Took 1.23 seconds to deallocate network for instance. [ 1120.086081] env[62952]: DEBUG nova.network.neutron [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.156762] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14a1262f-5433-42e0-9723-0fb07ef4098a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.169536] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c543079-0d00-429e-a99d-fb976f16dbeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.200393] env[62952]: DEBUG nova.compute.manager [req-cdca6912-515b-4744-b600-44a7ece9d4ab req-1e60b393-856a-46ed-97c6-c4712dabef27 service nova] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Detach interface failed, port_id=be83803d-9485-4459-882a-5dc9383e0f62, reason: Instance f5e5c5c1-e2af-40e0-a957-67b0fd22718d could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1120.298261] env[62952]: DEBUG nova.scheduler.client.report [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.433584] env[62952]: DEBUG nova.network.neutron [-] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.465681] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7f50b49-ef60-4531-84f5-4493254019c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.475460] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1526d72c-f840-4f95-a8da-3dba7a899432 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.500723] env[62952]: DEBUG nova.compute.manager [req-a1181df2-6023-4da3-958c-b2ed23013fd6 req-063d5c1d-0806-4992-9f82-037c5423a9c6 service nova] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Detach interface failed, port_id=6e4543b9-4601-4109-b4fc-14eee838c035, reason: Instance 75d640ed-c41a-4761-8867-191d8b3e1f79 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1120.572588] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.587022] env[62952]: INFO nova.compute.manager [-] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Took 1.50 seconds to deallocate network for instance. [ 1120.804774] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.807032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.235s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.808330] env[62952]: DEBUG nova.objects.instance [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lazy-loading 'resources' on Instance uuid 9036f863-9d77-491b-8aa9-978d0c400708 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.834350] env[62952]: INFO nova.scheduler.client.report [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance f302233b-2e1f-4c71-b000-0c9eb6530730 [ 1120.936398] env[62952]: INFO nova.compute.manager [-] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Took 1.48 seconds to deallocate network for instance. [ 1121.093674] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.341397] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c321270a-485c-48f3-9e7c-9f09a8d1776a tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "f302233b-2e1f-4c71-b000-0c9eb6530730" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.209s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.407027] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fca404-c330-4dad-ba19-7ff795de748a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.415062] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763ba9ee-e326-4ddf-9381-86a2c02f3b18 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.448424] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.449500] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fe1f22-f8a4-4f73-a8af-debddc7f7f88 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.457703] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9d6b79-1835-4e37-a33f-583789736f3d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.471466] env[62952]: DEBUG nova.compute.provider_tree [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.974684] env[62952]: DEBUG nova.scheduler.client.report [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.059245] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.059500] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.480127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.482635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.389s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.482921] env[62952]: DEBUG nova.objects.instance [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'resources' on Instance uuid f5e5c5c1-e2af-40e0-a957-67b0fd22718d {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.503507] env[62952]: INFO nova.scheduler.client.report [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Deleted allocations for instance 9036f863-9d77-491b-8aa9-978d0c400708 [ 1122.562863] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1123.011493] env[62952]: DEBUG oslo_concurrency.lockutils [None req-78f87fdd-f2e0-4ff7-a2df-e6f37b42429b tempest-ServerDiskConfigTestJSON-111952320 tempest-ServerDiskConfigTestJSON-111952320-project-member] Lock "9036f863-9d77-491b-8aa9-978d0c400708" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.288s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.087662] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.103057] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e230260a-3aa3-4819-9d93-e5773de82745 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.111740] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fefd0c-8d57-4b5a-a8c3-e9d8ac9d2722 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.140988] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1084e4f-3e68-4f30-8c6b-2717fd1812ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.148435] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f30c34-7418-42ed-b62d-050041fb3581 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.161948] env[62952]: DEBUG nova.compute.provider_tree [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.665268] env[62952]: DEBUG nova.scheduler.client.report [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.172022] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.173040] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.724s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.173040] env[62952]: DEBUG nova.objects.instance [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'resources' on Instance uuid 75d640ed-c41a-4761-8867-191d8b3e1f79 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.193140] env[62952]: INFO nova.scheduler.client.report [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted allocations for instance f5e5c5c1-e2af-40e0-a957-67b0fd22718d [ 1124.702980] env[62952]: DEBUG oslo_concurrency.lockutils [None req-697fe131-44d9-497c-b40e-33ff99fb9178 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "f5e5c5c1-e2af-40e0-a957-67b0fd22718d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.750s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.762036] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2303073-8bf3-4623-b5ec-24fef0a212c1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.770054] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fe2b32-fd8b-490c-a244-14a1c7ca2c0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.801805] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee17f6d-4f96-4f07-af9c-f64168a9fe31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.809407] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1418b4-16d5-4527-9adb-33035cc99ba6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.823305] env[62952]: DEBUG nova.compute.provider_tree [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.326711] env[62952]: DEBUG nova.scheduler.client.report [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.833150] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.835476] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.748s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.837102] env[62952]: INFO nova.compute.claims [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1125.857893] env[62952]: INFO nova.scheduler.client.report [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted allocations for instance 75d640ed-c41a-4761-8867-191d8b3e1f79 [ 1126.365587] env[62952]: DEBUG oslo_concurrency.lockutils [None req-069a3da6-ba36-42e2-ab49-dedf73fd79ec tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "75d640ed-c41a-4761-8867-191d8b3e1f79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.018s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.655889] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.656154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.927006] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78a0be4-707b-465b-9c96-6f42e77677ea {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.934650] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf25e66f-9ed5-49b4-99ba-0f0684801a72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.963056] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53766b5d-694e-4b38-bcf8-53c08b7ce7e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.970144] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76a4711-85b9-41a9-9ad6-f1d9934e5600 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.983204] env[62952]: DEBUG nova.compute.provider_tree [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.158507] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1127.486273] env[62952]: DEBUG nova.scheduler.client.report [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1127.676080] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.991479] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.156s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.992043] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.994780] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.319s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.996398] env[62952]: INFO nova.compute.claims [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1128.417331] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.417619] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.500709] env[62952]: DEBUG nova.compute.utils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1128.504081] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1128.504251] env[62952]: DEBUG nova.network.neutron [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1128.551180] env[62952]: DEBUG nova.policy [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7293e0f7b74c55aa0c67701bd7078e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b8e72818064a0f950614b39e30f717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1128.796069] env[62952]: DEBUG nova.network.neutron [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Successfully created port: 51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1128.919711] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1129.004995] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1129.110194] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902fb50f-68d2-48b9-9aa5-44cf6217aada {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.118226] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5c184a-0003-4994-bbd8-f6dd90b47acd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.147063] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e130e2-45bf-47b5-a960-f769c593b3d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.153841] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efaca4b6-e3da-4ee4-a369-70cb937b4a3e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.167706] env[62952]: DEBUG nova.compute.provider_tree [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.439048] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.671305] env[62952]: DEBUG nova.scheduler.client.report [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.018462] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1130.043330] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1130.043611] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1130.043843] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1130.044117] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1130.044317] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1130.044509] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1130.044760] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1130.044971] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1130.045181] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1130.045390] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1130.045614] env[62952]: DEBUG nova.virt.hardware [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1130.046632] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da66711-888e-440d-afc8-60a1000f2b7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.055589] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1ab2f0-e44d-41dc-89a5-9868d6c09de7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.176070] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.176609] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1130.182036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.740s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.182036] env[62952]: INFO nova.compute.claims [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1130.247089] env[62952]: DEBUG nova.compute.manager [req-0acd0fbe-bd4b-426e-838f-4731d6f0c89f req-bb8ea82d-256e-4325-b029-c7e67201c276 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Received event network-vif-plugged-51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.247358] env[62952]: DEBUG oslo_concurrency.lockutils [req-0acd0fbe-bd4b-426e-838f-4731d6f0c89f req-bb8ea82d-256e-4325-b029-c7e67201c276 service nova] Acquiring lock "2fa96864-5440-454c-a743-fdf4e70a0e2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.247573] env[62952]: DEBUG oslo_concurrency.lockutils [req-0acd0fbe-bd4b-426e-838f-4731d6f0c89f req-bb8ea82d-256e-4325-b029-c7e67201c276 service nova] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.247747] env[62952]: DEBUG oslo_concurrency.lockutils [req-0acd0fbe-bd4b-426e-838f-4731d6f0c89f req-bb8ea82d-256e-4325-b029-c7e67201c276 service nova] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.247919] env[62952]: DEBUG nova.compute.manager [req-0acd0fbe-bd4b-426e-838f-4731d6f0c89f req-bb8ea82d-256e-4325-b029-c7e67201c276 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] No waiting events found dispatching network-vif-plugged-51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1130.249675] env[62952]: WARNING nova.compute.manager [req-0acd0fbe-bd4b-426e-838f-4731d6f0c89f req-bb8ea82d-256e-4325-b029-c7e67201c276 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Received unexpected event network-vif-plugged-51053d56-b119-40ed-ab17-c1cb0f94aa79 for instance with vm_state building and task_state spawning. [ 1130.342682] env[62952]: DEBUG nova.network.neutron [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Successfully updated port: 51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1130.686157] env[62952]: DEBUG nova.compute.utils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1130.689704] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1130.689704] env[62952]: DEBUG nova.network.neutron [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1130.726662] env[62952]: DEBUG nova.policy [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '913eacf6a94845b189e0bd5df408138a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8c2a296e8fe4805b9bb761f91a407f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1130.845945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "refresh_cache-2fa96864-5440-454c-a743-fdf4e70a0e2e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.846118] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "refresh_cache-2fa96864-5440-454c-a743-fdf4e70a0e2e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.846265] env[62952]: DEBUG nova.network.neutron [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1130.947635] env[62952]: DEBUG nova.network.neutron [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Successfully created port: 65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1131.191249] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1131.288978] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a064a5-d59b-485e-9a74-65b237735194 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.298074] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd13454-5101-4607-a710-dcbe25a1b767 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.326668] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5783f2-ab73-4851-ab63-f8f1400afb9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.333114] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e72f42b-5e01-43ad-80ca-587cf4f34300 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.345365] env[62952]: DEBUG nova.compute.provider_tree [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.394721] env[62952]: DEBUG nova.network.neutron [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1131.559872] env[62952]: DEBUG nova.network.neutron [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Updating instance_info_cache with network_info: [{"id": "51053d56-b119-40ed-ab17-c1cb0f94aa79", "address": "fa:16:3e:f7:03:36", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51053d56-b1", "ovs_interfaceid": "51053d56-b119-40ed-ab17-c1cb0f94aa79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.848263] env[62952]: DEBUG nova.scheduler.client.report [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.062304] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "refresh_cache-2fa96864-5440-454c-a743-fdf4e70a0e2e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.062633] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Instance network_info: |[{"id": "51053d56-b119-40ed-ab17-c1cb0f94aa79", "address": "fa:16:3e:f7:03:36", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51053d56-b1", "ovs_interfaceid": "51053d56-b119-40ed-ab17-c1cb0f94aa79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1132.063075] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:03:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51053d56-b119-40ed-ab17-c1cb0f94aa79', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1132.071463] env[62952]: DEBUG oslo.service.loopingcall [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.071731] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1132.071974] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e07178e-8c76-4a75-89fd-4ffc00e91298 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.094726] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.094726] env[62952]: value = "task-1367701" [ 1132.094726] env[62952]: _type = "Task" [ 1132.094726] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.102829] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.202158] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1132.231634] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1132.231964] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1132.232150] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.232342] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1132.232495] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.232672] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1132.232975] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1132.233165] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1132.233351] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1132.233522] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1132.233703] env[62952]: DEBUG nova.virt.hardware [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1132.234671] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44fe497-010b-4dfd-bb81-03c69a260e4d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.244799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba037920-ea39-46f1-9e1b-b29c67f74417 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.353819] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.174s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.354243] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1132.396480] env[62952]: DEBUG nova.compute.manager [req-b51607af-ec8e-46d6-bfcc-563c2c3efced req-acc0cffc-b0eb-4ff8-a70f-03a6cb081e56 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Received event network-vif-plugged-65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1132.396717] env[62952]: DEBUG oslo_concurrency.lockutils [req-b51607af-ec8e-46d6-bfcc-563c2c3efced req-acc0cffc-b0eb-4ff8-a70f-03a6cb081e56 service nova] Acquiring lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.397016] env[62952]: DEBUG oslo_concurrency.lockutils [req-b51607af-ec8e-46d6-bfcc-563c2c3efced req-acc0cffc-b0eb-4ff8-a70f-03a6cb081e56 service nova] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.397123] env[62952]: DEBUG oslo_concurrency.lockutils [req-b51607af-ec8e-46d6-bfcc-563c2c3efced req-acc0cffc-b0eb-4ff8-a70f-03a6cb081e56 service nova] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.397286] env[62952]: DEBUG nova.compute.manager [req-b51607af-ec8e-46d6-bfcc-563c2c3efced req-acc0cffc-b0eb-4ff8-a70f-03a6cb081e56 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] No waiting events found dispatching network-vif-plugged-65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1132.397452] env[62952]: WARNING nova.compute.manager [req-b51607af-ec8e-46d6-bfcc-563c2c3efced req-acc0cffc-b0eb-4ff8-a70f-03a6cb081e56 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Received unexpected event network-vif-plugged-65c2ecad-019b-4edb-a82c-3cc676089eab for instance with vm_state building and task_state spawning. [ 1132.444791] env[62952]: DEBUG nova.compute.manager [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Received event network-changed-51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1132.444924] env[62952]: DEBUG nova.compute.manager [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Refreshing instance network info cache due to event network-changed-51053d56-b119-40ed-ab17-c1cb0f94aa79. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1132.445325] env[62952]: DEBUG oslo_concurrency.lockutils [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] Acquiring lock "refresh_cache-2fa96864-5440-454c-a743-fdf4e70a0e2e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.445325] env[62952]: DEBUG oslo_concurrency.lockutils [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] Acquired lock "refresh_cache-2fa96864-5440-454c-a743-fdf4e70a0e2e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.445431] env[62952]: DEBUG nova.network.neutron [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Refreshing network info cache for port 51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1132.487642] env[62952]: DEBUG nova.network.neutron [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Successfully updated port: 65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1132.604818] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.859483] env[62952]: DEBUG nova.compute.utils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1132.860859] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1132.861036] env[62952]: DEBUG nova.network.neutron [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1132.899842] env[62952]: DEBUG nova.policy [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e67193d02201461bb7e4339b5025ca48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfc390d64c0463190f071f3f62936dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1132.990150] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.990309] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.990537] env[62952]: DEBUG nova.network.neutron [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1133.105694] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.151702] env[62952]: DEBUG nova.network.neutron [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Successfully created port: c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1133.161432] env[62952]: DEBUG nova.network.neutron [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Updated VIF entry in instance network info cache for port 51053d56-b119-40ed-ab17-c1cb0f94aa79. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1133.161432] env[62952]: DEBUG nova.network.neutron [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Updating instance_info_cache with network_info: [{"id": "51053d56-b119-40ed-ab17-c1cb0f94aa79", "address": "fa:16:3e:f7:03:36", "network": {"id": "c5c2d590-aa7e-43c1-9689-c2a007c9502f", "bridge": "br-int", "label": "tempest-ServersTestJSON-117146209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30b8e72818064a0f950614b39e30f717", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51053d56-b1", "ovs_interfaceid": "51053d56-b119-40ed-ab17-c1cb0f94aa79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.363864] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1133.521284] env[62952]: DEBUG nova.network.neutron [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1133.607703] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.644617] env[62952]: DEBUG nova.network.neutron [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updating instance_info_cache with network_info: [{"id": "65c2ecad-019b-4edb-a82c-3cc676089eab", "address": "fa:16:3e:f5:06:67", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c2ecad-01", "ovs_interfaceid": "65c2ecad-019b-4edb-a82c-3cc676089eab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.663296] env[62952]: DEBUG oslo_concurrency.lockutils [req-666a20cb-1687-4eb7-84ac-87da169ac1ce req-3a121b83-31e2-4a04-9471-288533292079 service nova] Releasing lock "refresh_cache-2fa96864-5440-454c-a743-fdf4e70a0e2e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.106848] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.147593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.147885] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Instance network_info: |[{"id": "65c2ecad-019b-4edb-a82c-3cc676089eab", "address": "fa:16:3e:f5:06:67", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c2ecad-01", "ovs_interfaceid": "65c2ecad-019b-4edb-a82c-3cc676089eab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1134.148344] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:06:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65c2ecad-019b-4edb-a82c-3cc676089eab', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1134.155774] env[62952]: DEBUG oslo.service.loopingcall [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.155989] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1134.156224] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f67d2471-1130-4a03-8824-10bc4fae99d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.176389] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1134.176389] env[62952]: value = "task-1367702" [ 1134.176389] env[62952]: _type = "Task" [ 1134.176389] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.183707] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367702, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.373841] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1134.401720] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1134.401973] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1134.402147] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1134.402333] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1134.402484] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1134.402636] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1134.402853] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1134.403010] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1134.403189] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1134.403355] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1134.403532] env[62952]: DEBUG nova.virt.hardware [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1134.404437] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea028eb-32c8-454f-a684-550bc0395b42 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.412273] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a1c12e-c4b3-42fc-9c9c-ee2c059ef31e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.472738] env[62952]: DEBUG nova.compute.manager [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Received event network-changed-65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.472963] env[62952]: DEBUG nova.compute.manager [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Refreshing instance network info cache due to event network-changed-65c2ecad-019b-4edb-a82c-3cc676089eab. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1134.473159] env[62952]: DEBUG oslo_concurrency.lockutils [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] Acquiring lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.473316] env[62952]: DEBUG oslo_concurrency.lockutils [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] Acquired lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.473486] env[62952]: DEBUG nova.network.neutron [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Refreshing network info cache for port 65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1134.607786] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.629542] env[62952]: DEBUG nova.network.neutron [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Successfully updated port: c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1134.685732] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367702, 'name': CreateVM_Task, 'duration_secs': 0.311321} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.687124] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1134.687124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.687124] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.687414] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1134.687720] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ff3380d-3c54-404f-9500-42f8225c92ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.691954] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1134.691954] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a107c1-f704-e3c1-6a94-ea8bf4752d92" [ 1134.691954] env[62952]: _type = "Task" [ 1134.691954] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.699348] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a107c1-f704-e3c1-6a94-ea8bf4752d92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.109298] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.132344] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.132488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.132658] env[62952]: DEBUG nova.network.neutron [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1135.173038] env[62952]: DEBUG nova.network.neutron [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updated VIF entry in instance network info cache for port 65c2ecad-019b-4edb-a82c-3cc676089eab. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1135.173419] env[62952]: DEBUG nova.network.neutron [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updating instance_info_cache with network_info: [{"id": "65c2ecad-019b-4edb-a82c-3cc676089eab", "address": "fa:16:3e:f5:06:67", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c2ecad-01", "ovs_interfaceid": "65c2ecad-019b-4edb-a82c-3cc676089eab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.201525] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a107c1-f704-e3c1-6a94-ea8bf4752d92, 'name': SearchDatastore_Task, 'duration_secs': 0.023041} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.201839] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.202083] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1135.202326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.202480] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.202663] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1135.202915] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23ee6623-fc39-45fc-bb03-b72ae18c29e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.211025] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1135.211217] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1135.211960] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e10a973-f494-4056-b206-20c58f528d75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.216768] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1135.216768] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52627b0f-6084-3352-3d5d-557d05103650" [ 1135.216768] env[62952]: _type = "Task" [ 1135.216768] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.224418] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52627b0f-6084-3352-3d5d-557d05103650, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.608566] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.674986] env[62952]: DEBUG nova.network.neutron [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1135.677092] env[62952]: DEBUG oslo_concurrency.lockutils [req-b4ccdf7c-51a1-4cdf-b0c7-e79ee517f0c0 req-0ec2cdf1-3c0b-49c8-8355-b98b49e47676 service nova] Releasing lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.727098] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52627b0f-6084-3352-3d5d-557d05103650, 'name': SearchDatastore_Task, 'duration_secs': 0.008122} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.730341] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3895ebf3-9f7b-4074-b2b5-719de6e16f5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.735689] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1135.735689] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52277ba6-abdf-af05-7d9b-ede88efed132" [ 1135.735689] env[62952]: _type = "Task" [ 1135.735689] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.743649] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52277ba6-abdf-af05-7d9b-ede88efed132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.821357] env[62952]: DEBUG nova.network.neutron [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updating instance_info_cache with network_info: [{"id": "c00369d7-037b-4a8e-9f3e-cd130492b005", "address": "fa:16:3e:d7:06:5d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc00369d7-03", "ovs_interfaceid": "c00369d7-037b-4a8e-9f3e-cd130492b005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.110908] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.245767] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52277ba6-abdf-af05-7d9b-ede88efed132, 'name': SearchDatastore_Task, 'duration_secs': 0.009699} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.246050] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.246319] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5/3cc1f262-3e94-4a7e-997c-fbd24bad0aa5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1136.246574] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9b1fa65-2d92-488b-a121-eeb431d55d70 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.252914] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1136.252914] env[62952]: value = "task-1367703" [ 1136.252914] env[62952]: _type = "Task" [ 1136.252914] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.260264] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.324177] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.324520] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Instance network_info: |[{"id": "c00369d7-037b-4a8e-9f3e-cd130492b005", "address": "fa:16:3e:d7:06:5d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc00369d7-03", "ovs_interfaceid": "c00369d7-037b-4a8e-9f3e-cd130492b005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1136.325049] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:06:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c00369d7-037b-4a8e-9f3e-cd130492b005', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1136.333248] env[62952]: DEBUG oslo.service.loopingcall [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.333525] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1136.333790] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-384c3727-75c1-4ac2-a32a-fe91411578ca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.353974] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1136.353974] env[62952]: value = "task-1367704" [ 1136.353974] env[62952]: _type = "Task" [ 1136.353974] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.362630] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367704, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.505128] env[62952]: DEBUG nova.compute.manager [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Received event network-vif-plugged-c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1136.505383] env[62952]: DEBUG oslo_concurrency.lockutils [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.505623] env[62952]: DEBUG oslo_concurrency.lockutils [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.505818] env[62952]: DEBUG oslo_concurrency.lockutils [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.506015] env[62952]: DEBUG nova.compute.manager [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] No waiting events found dispatching network-vif-plugged-c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1136.506205] env[62952]: WARNING nova.compute.manager [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Received unexpected event network-vif-plugged-c00369d7-037b-4a8e-9f3e-cd130492b005 for instance with vm_state building and task_state spawning. [ 1136.506439] env[62952]: DEBUG nova.compute.manager [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Received event network-changed-c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1136.506635] env[62952]: DEBUG nova.compute.manager [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Refreshing instance network info cache due to event network-changed-c00369d7-037b-4a8e-9f3e-cd130492b005. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1136.506854] env[62952]: DEBUG oslo_concurrency.lockutils [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] Acquiring lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.507033] env[62952]: DEBUG oslo_concurrency.lockutils [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] Acquired lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.507207] env[62952]: DEBUG nova.network.neutron [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Refreshing network info cache for port c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1136.612412] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.764898] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438578} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.765203] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5/3cc1f262-3e94-4a7e-997c-fbd24bad0aa5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1136.765427] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1136.765691] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9297c90-5a64-4e66-91e6-a1c9e7dd075e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.772410] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1136.772410] env[62952]: value = "task-1367705" [ 1136.772410] env[62952]: _type = "Task" [ 1136.772410] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.781504] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367705, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.865962] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367704, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.109452] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367701, 'name': CreateVM_Task, 'duration_secs': 4.701911} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.109614] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1137.110284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.110488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.110821] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1137.111079] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df4a4d2d-2882-4864-8247-6a14e4e380b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.115449] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1137.115449] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a8e9ca-41b0-6502-df98-b1e89c3a8f4d" [ 1137.115449] env[62952]: _type = "Task" [ 1137.115449] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.122636] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a8e9ca-41b0-6502-df98-b1e89c3a8f4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.187858] env[62952]: DEBUG nova.network.neutron [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updated VIF entry in instance network info cache for port c00369d7-037b-4a8e-9f3e-cd130492b005. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1137.188238] env[62952]: DEBUG nova.network.neutron [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updating instance_info_cache with network_info: [{"id": "c00369d7-037b-4a8e-9f3e-cd130492b005", "address": "fa:16:3e:d7:06:5d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc00369d7-03", "ovs_interfaceid": "c00369d7-037b-4a8e-9f3e-cd130492b005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.282581] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367705, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084182} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.282804] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1137.283845] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0433d10c-9402-4062-91b5-f96528781b7f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.305716] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5/3cc1f262-3e94-4a7e-997c-fbd24bad0aa5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.305940] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37d86194-fe07-4b33-b18b-03fb7f92a1dc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.324849] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1137.324849] env[62952]: value = "task-1367706" [ 1137.324849] env[62952]: _type = "Task" [ 1137.324849] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.332344] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367706, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.363559] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367704, 'name': CreateVM_Task, 'duration_secs': 0.543755} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.363734] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1137.364434] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.626253] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a8e9ca-41b0-6502-df98-b1e89c3a8f4d, 'name': SearchDatastore_Task, 'duration_secs': 0.012328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.626626] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.626762] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1137.627013] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.627175] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.627357] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1137.627639] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.627941] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1137.628180] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-892eb0ff-f4f8-45e6-abad-31082250e33f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.629902] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e35cede-9f33-498b-81e6-f133e15c259b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.634417] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1137.634417] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52d9fc78-6136-ccbd-c6cc-f8c4740bd50e" [ 1137.634417] env[62952]: _type = "Task" [ 1137.634417] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.641727] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1137.641906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1137.645124] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15a9ad3a-d3e3-4338-b948-0e4db42176eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.647092] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d9fc78-6136-ccbd-c6cc-f8c4740bd50e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.649858] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1137.649858] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528054e1-bf24-6385-2fca-5674bca2a75a" [ 1137.649858] env[62952]: _type = "Task" [ 1137.649858] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.656834] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528054e1-bf24-6385-2fca-5674bca2a75a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.690620] env[62952]: DEBUG oslo_concurrency.lockutils [req-1991bd59-ec90-48e7-8466-1ebad10ed3e2 req-8468837d-7ace-476e-be96-c798001ac6f8 service nova] Releasing lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.834259] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367706, 'name': ReconfigVM_Task, 'duration_secs': 0.299071} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.834525] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5/3cc1f262-3e94-4a7e-997c-fbd24bad0aa5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.835144] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f16cc4dd-4805-4cde-a39c-ad1291ce52a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.841096] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1137.841096] env[62952]: value = "task-1367707" [ 1137.841096] env[62952]: _type = "Task" [ 1137.841096] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.849381] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367707, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.144171] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52d9fc78-6136-ccbd-c6cc-f8c4740bd50e, 'name': SearchDatastore_Task, 'duration_secs': 0.017194} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.144471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.144709] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1138.144926] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.158952] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]528054e1-bf24-6385-2fca-5674bca2a75a, 'name': SearchDatastore_Task, 'duration_secs': 0.008141} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.159299] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7a24031-bee7-4d54-807c-af3e8560a5a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.163709] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1138.163709] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]522bbb0f-8ac5-f433-f1e9-7dd9dbedc6de" [ 1138.163709] env[62952]: _type = "Task" [ 1138.163709] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.170877] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522bbb0f-8ac5-f433-f1e9-7dd9dbedc6de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.350300] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367707, 'name': Rename_Task, 'duration_secs': 0.150677} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.351062] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1138.351168] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ed02fe9-1bec-4458-b7f8-2301d4ebbd52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.357237] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1138.357237] env[62952]: value = "task-1367708" [ 1138.357237] env[62952]: _type = "Task" [ 1138.357237] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.364174] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367708, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.673601] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]522bbb0f-8ac5-f433-f1e9-7dd9dbedc6de, 'name': SearchDatastore_Task, 'duration_secs': 0.009456} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.673935] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.674137] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 2fa96864-5440-454c-a743-fdf4e70a0e2e/2fa96864-5440-454c-a743-fdf4e70a0e2e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1138.674456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.674722] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1138.674960] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-554fe66e-6e66-4b24-a028-6c717a2c10fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.677027] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0fa342b-8045-450f-be07-36a7f837ddf2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.684118] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1138.684118] env[62952]: value = "task-1367709" [ 1138.684118] env[62952]: _type = "Task" [ 1138.684118] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.687861] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1138.688052] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1138.688760] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0af265a5-67f4-4359-b658-247fcb2ca13a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.695659] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.696917] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1138.696917] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]523dfec5-8a00-5b1a-cf43-d28d1d39a66c" [ 1138.696917] env[62952]: _type = "Task" [ 1138.696917] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.704872] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]523dfec5-8a00-5b1a-cf43-d28d1d39a66c, 'name': SearchDatastore_Task, 'duration_secs': 0.008663} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.705674] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb0f767e-79d5-4d11-bf4a-241930b8394d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.710229] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1138.710229] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52a3f7cc-84e8-5e1d-0feb-b36f43bd7375" [ 1138.710229] env[62952]: _type = "Task" [ 1138.710229] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.717290] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a3f7cc-84e8-5e1d-0feb-b36f43bd7375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.867525] env[62952]: DEBUG oslo_vmware.api [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367708, 'name': PowerOnVM_Task, 'duration_secs': 0.43298} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.867873] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1138.868134] env[62952]: INFO nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1138.868367] env[62952]: DEBUG nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.869215] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35512ed-ebd9-4faa-ae76-384390f6a3a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.195731] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453737} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.195975] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 2fa96864-5440-454c-a743-fdf4e70a0e2e/2fa96864-5440-454c-a743-fdf4e70a0e2e.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1139.196214] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1139.196477] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75085b1e-048e-41bd-b961-6d86280705b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.206171] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1139.206171] env[62952]: value = "task-1367710" [ 1139.206171] env[62952]: _type = "Task" [ 1139.206171] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.213647] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367710, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.221675] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52a3f7cc-84e8-5e1d-0feb-b36f43bd7375, 'name': SearchDatastore_Task, 'duration_secs': 0.007849} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.222093] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.222199] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 26fdfac2-4ea1-4822-b38c-d8f7388436be/26fdfac2-4ea1-4822-b38c-d8f7388436be.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1139.222460] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-344d0151-49d8-4dca-8ba2-ab7092f3ccb9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.229305] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1139.229305] env[62952]: value = "task-1367711" [ 1139.229305] env[62952]: _type = "Task" [ 1139.229305] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.238945] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.388789] env[62952]: INFO nova.compute.manager [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Took 11.73 seconds to build instance. [ 1139.717670] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367710, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07343} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.717670] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1139.717670] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54d431c-ff19-40f9-912d-712085932951 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.738833] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 2fa96864-5440-454c-a743-fdf4e70a0e2e/2fa96864-5440-454c-a743-fdf4e70a0e2e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.741973] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03c25b7f-fcbd-49da-9986-dab85fc6bc05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.762897] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477565} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.765382] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore1] 26fdfac2-4ea1-4822-b38c-d8f7388436be/26fdfac2-4ea1-4822-b38c-d8f7388436be.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1139.765382] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1139.765382] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1139.765382] env[62952]: value = "task-1367712" [ 1139.765382] env[62952]: _type = "Task" [ 1139.765382] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.765382] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5481e7d-2506-43a9-aacf-7fcfe5ecb818 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.774718] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367712, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.775889] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1139.775889] env[62952]: value = "task-1367713" [ 1139.775889] env[62952]: _type = "Task" [ 1139.775889] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.783688] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367713, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.786128] env[62952]: DEBUG nova.compute.manager [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Received event network-changed-65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.786291] env[62952]: DEBUG nova.compute.manager [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Refreshing instance network info cache due to event network-changed-65c2ecad-019b-4edb-a82c-3cc676089eab. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1139.786509] env[62952]: DEBUG oslo_concurrency.lockutils [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] Acquiring lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.786674] env[62952]: DEBUG oslo_concurrency.lockutils [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] Acquired lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.786849] env[62952]: DEBUG nova.network.neutron [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Refreshing network info cache for port 65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1139.891413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3ff6b457-fea3-4fd8-89df-33f51ac14c78 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.235s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.277050] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367712, 'name': ReconfigVM_Task, 'duration_secs': 0.295731} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.280070] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 2fa96864-5440-454c-a743-fdf4e70a0e2e/2fa96864-5440-454c-a743-fdf4e70a0e2e.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1140.280716] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb12bb00-5780-4dc1-b070-da69839eb952 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.287311] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367713, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06889} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.288416] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1140.288741] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1140.288741] env[62952]: value = "task-1367714" [ 1140.288741] env[62952]: _type = "Task" [ 1140.288741] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.289407] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc1ef6a-6173-49df-ad94-a1be0772dbc7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.315434] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 26fdfac2-4ea1-4822-b38c-d8f7388436be/26fdfac2-4ea1-4822-b38c-d8f7388436be.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1140.319022] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19168c36-41fc-4bd4-9789-752ca268ffb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.335756] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367714, 'name': Rename_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.344033] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1140.344033] env[62952]: value = "task-1367715" [ 1140.344033] env[62952]: _type = "Task" [ 1140.344033] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.351961] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367715, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.505468] env[62952]: DEBUG nova.network.neutron [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updated VIF entry in instance network info cache for port 65c2ecad-019b-4edb-a82c-3cc676089eab. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1140.505833] env[62952]: DEBUG nova.network.neutron [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updating instance_info_cache with network_info: [{"id": "65c2ecad-019b-4edb-a82c-3cc676089eab", "address": "fa:16:3e:f5:06:67", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c2ecad-01", "ovs_interfaceid": "65c2ecad-019b-4edb-a82c-3cc676089eab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.802545] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367714, 'name': Rename_Task, 'duration_secs': 0.165687} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.802850] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1140.803057] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-393388fa-74fd-4b42-b472-8d9e1dab7283 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.809425] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1140.809425] env[62952]: value = "task-1367716" [ 1140.809425] env[62952]: _type = "Task" [ 1140.809425] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.816537] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.852290] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367715, 'name': ReconfigVM_Task, 'duration_secs': 0.291697} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.852594] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 26fdfac2-4ea1-4822-b38c-d8f7388436be/26fdfac2-4ea1-4822-b38c-d8f7388436be.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1140.853357] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d1c4e90-343d-4d14-9b2c-aad44cc79ef4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.859272] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1140.859272] env[62952]: value = "task-1367718" [ 1140.859272] env[62952]: _type = "Task" [ 1140.859272] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.866642] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367718, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.008738] env[62952]: DEBUG oslo_concurrency.lockutils [req-4df57fb8-0024-4c57-aa62-3663214227a7 req-2b7902f4-2a06-4793-8938-a3ca5d9d7b44 service nova] Releasing lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.319822] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367716, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.369087] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367718, 'name': Rename_Task, 'duration_secs': 0.179251} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.369384] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1141.369634] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ce07df1-33ad-4e09-b717-11c9c0afab76 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.375623] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1141.375623] env[62952]: value = "task-1367719" [ 1141.375623] env[62952]: _type = "Task" [ 1141.375623] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.388013] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.820321] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367716, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.885992] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367719, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.320987] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367716, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.385138] env[62952]: DEBUG oslo_vmware.api [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367719, 'name': PowerOnVM_Task, 'duration_secs': 0.742382} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.385434] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1142.385747] env[62952]: INFO nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Took 8.01 seconds to spawn the instance on the hypervisor. [ 1142.385814] env[62952]: DEBUG nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1142.386634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffa1b9a-0601-4fe3-b435-81d49a4b59a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.824480] env[62952]: DEBUG oslo_vmware.api [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367716, 'name': PowerOnVM_Task, 'duration_secs': 1.660637} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.824893] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1142.825072] env[62952]: INFO nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Took 12.81 seconds to spawn the instance on the hypervisor. [ 1142.825306] env[62952]: DEBUG nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1142.826148] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574d5a46-dc4b-459b-86c9-06f9b5607ecf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.902509] env[62952]: INFO nova.compute.manager [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Took 13.48 seconds to build instance. [ 1143.200221] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.200422] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.342071] env[62952]: INFO nova.compute.manager [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Took 20.27 seconds to build instance. [ 1143.404257] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b7f1c836-668d-4d1b-b7b3-dcccc3439dc3 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.986s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.704080] env[62952]: DEBUG nova.compute.utils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.843355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2a2fc4f-5fa9-408a-808d-eb83398ef931 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.784s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.907741] env[62952]: DEBUG nova.compute.manager [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Received event network-changed-c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1143.908061] env[62952]: DEBUG nova.compute.manager [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Refreshing instance network info cache due to event network-changed-c00369d7-037b-4a8e-9f3e-cd130492b005. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1143.908629] env[62952]: DEBUG oslo_concurrency.lockutils [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] Acquiring lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.908629] env[62952]: DEBUG oslo_concurrency.lockutils [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] Acquired lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.908629] env[62952]: DEBUG nova.network.neutron [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Refreshing network info cache for port c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1144.209038] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.388532] env[62952]: DEBUG oslo_concurrency.lockutils [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.388821] env[62952]: DEBUG oslo_concurrency.lockutils [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.389050] env[62952]: DEBUG nova.compute.manager [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1144.390285] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24be890-5716-484a-a0af-cd5c1c63a96c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.397396] env[62952]: DEBUG nova.compute.manager [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62952) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1144.397962] env[62952]: DEBUG nova.objects.instance [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'flavor' on Instance uuid 2fa96864-5440-454c-a743-fdf4e70a0e2e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.628322] env[62952]: DEBUG nova.network.neutron [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updated VIF entry in instance network info cache for port c00369d7-037b-4a8e-9f3e-cd130492b005. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1144.628754] env[62952]: DEBUG nova.network.neutron [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updating instance_info_cache with network_info: [{"id": "c00369d7-037b-4a8e-9f3e-cd130492b005", "address": "fa:16:3e:d7:06:5d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc00369d7-03", "ovs_interfaceid": "c00369d7-037b-4a8e-9f3e-cd130492b005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.903650] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1144.904126] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f42fe09-36d8-4e05-8062-e268a71f8db8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.912179] env[62952]: DEBUG oslo_vmware.api [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1144.912179] env[62952]: value = "task-1367721" [ 1144.912179] env[62952]: _type = "Task" [ 1144.912179] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.920341] env[62952]: DEBUG oslo_vmware.api [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.132775] env[62952]: DEBUG oslo_concurrency.lockutils [req-79590158-5d4e-4706-ad3c-6fbfe65e3d65 req-3358cf33-11e2-4761-9faa-bfe9bcf34027 service nova] Releasing lock "refresh_cache-26fdfac2-4ea1-4822-b38c-d8f7388436be" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.269053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.269314] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.269567] env[62952]: INFO nova.compute.manager [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Attaching volume 16307e9b-ae51-423e-b715-36fca8845510 to /dev/sdb [ 1145.302038] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93453462-54f0-4df1-97b8-5c083ed7e9c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.309826] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75840c3d-9868-4dc2-a3e0-a26b4230eaf6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.323084] env[62952]: DEBUG nova.virt.block_device [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updating existing volume attachment record: 0008a41a-88d3-44bd-a1a1-32a8cbaedda7 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1145.421686] env[62952]: DEBUG oslo_vmware.api [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367721, 'name': PowerOffVM_Task, 'duration_secs': 0.417969} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.421954] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1145.422158] env[62952]: DEBUG nova.compute.manager [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1145.422908] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2758df5e-4424-4361-8b2d-88e11961cee2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.934477] env[62952]: DEBUG oslo_concurrency.lockutils [None req-837aa657-1c73-4e9e-aa28-eb8cfb982e38 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.078471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.078782] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.078983] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "2fa96864-5440-454c-a743-fdf4e70a0e2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.079209] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.079390] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.081571] env[62952]: INFO nova.compute.manager [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Terminating instance [ 1147.083285] env[62952]: DEBUG nova.compute.manager [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1147.083487] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1147.084319] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987c6251-7c86-4071-9956-3cc3bd2d6364 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.091877] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1147.092116] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc7f935d-dcbb-4b1f-a2de-162582438a72 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.169009] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1147.169250] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1147.169437] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore1] 2fa96864-5440-454c-a743-fdf4e70a0e2e {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.169697] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8366a604-9c19-4d09-8fb9-6e46985ff87f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.176573] env[62952]: DEBUG oslo_vmware.api [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1147.176573] env[62952]: value = "task-1367725" [ 1147.176573] env[62952]: _type = "Task" [ 1147.176573] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.184196] env[62952]: DEBUG oslo_vmware.api [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367725, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.686878] env[62952]: DEBUG oslo_vmware.api [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367725, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.454139} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.687165] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.687353] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1147.687535] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1147.687721] env[62952]: INFO nova.compute.manager [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1147.687965] env[62952]: DEBUG oslo.service.loopingcall [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.688183] env[62952]: DEBUG nova.compute.manager [-] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1147.688276] env[62952]: DEBUG nova.network.neutron [-] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1147.714662] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.714911] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.953385] env[62952]: DEBUG nova.compute.manager [req-7d191916-b5f5-48e5-b397-802dcda9d44f req-e533cdc7-4eea-4d31-a603-e69f0e3e05ee service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Received event network-vif-deleted-51053d56-b119-40ed-ab17-c1cb0f94aa79 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.953469] env[62952]: INFO nova.compute.manager [req-7d191916-b5f5-48e5-b397-802dcda9d44f req-e533cdc7-4eea-4d31-a603-e69f0e3e05ee service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Neutron deleted interface 51053d56-b119-40ed-ab17-c1cb0f94aa79; detaching it from the instance and deleting it from the info cache [ 1147.953622] env[62952]: DEBUG nova.network.neutron [req-7d191916-b5f5-48e5-b397-802dcda9d44f req-e533cdc7-4eea-4d31-a603-e69f0e3e05ee service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.221219] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1148.221219] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1148.221219] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1148.430440] env[62952]: DEBUG nova.network.neutron [-] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.455676] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92434ffb-a744-4cf7-8b14-7cbeeb14e975 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.465938] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33562129-9957-4ab6-ae96-4d6a6ad301b4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.492084] env[62952]: DEBUG nova.compute.manager [req-7d191916-b5f5-48e5-b397-802dcda9d44f req-e533cdc7-4eea-4d31-a603-e69f0e3e05ee service nova] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Detach interface failed, port_id=51053d56-b119-40ed-ab17-c1cb0f94aa79, reason: Instance 2fa96864-5440-454c-a743-fdf4e70a0e2e could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1148.726248] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Skipping network cache update for instance because it is being deleted. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1148.751777] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.751929] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.752086] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1148.752243] env[62952]: DEBUG nova.objects.instance [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lazy-loading 'info_cache' on Instance uuid aef28168-98a7-4f65-80e7-731633339abf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.932555] env[62952]: INFO nova.compute.manager [-] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Took 1.24 seconds to deallocate network for instance. [ 1149.281235] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.281523] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.439471] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.439802] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.440055] env[62952]: DEBUG nova.objects.instance [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid 2fa96864-5440-454c-a743-fdf4e70a0e2e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.785750] env[62952]: DEBUG nova.compute.utils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1149.865885] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1149.866152] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290988', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'name': 'volume-16307e9b-ae51-423e-b715-36fca8845510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'serial': '16307e9b-ae51-423e-b715-36fca8845510'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1149.867048] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b4a41d-e9c5-42be-9115-e2cca3eed1a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.885881] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff4d4dd-b970-4c64-8aed-0a6de487cea9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.910596] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-16307e9b-ae51-423e-b715-36fca8845510/volume-16307e9b-ae51-423e-b715-36fca8845510.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.910957] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2cd2bc5-351c-49d5-9c3f-85452356c275 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.929197] env[62952]: DEBUG oslo_vmware.api [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1149.929197] env[62952]: value = "task-1367728" [ 1149.929197] env[62952]: _type = "Task" [ 1149.929197] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.937279] env[62952]: DEBUG oslo_vmware.api [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367728, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.155556] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04dfde39-ff2a-4b4f-a09c-7487adf8e01a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.163342] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652329e0-85ee-47ad-94b4-fa85ea46ce4a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.192135] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433a1137-682a-4eed-9a3b-6ae5c776355d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.199368] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926d0e84-23f4-4db1-8c6b-9180de5f037f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.212184] env[62952]: DEBUG nova.compute.provider_tree [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.287962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.441294] env[62952]: DEBUG oslo_vmware.api [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367728, 'name': ReconfigVM_Task, 'duration_secs': 0.409253} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.441580] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-16307e9b-ae51-423e-b715-36fca8845510/volume-16307e9b-ae51-423e-b715-36fca8845510.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.446223] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc751590-8bb1-4685-b28d-db8522e5f274 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.460298] env[62952]: DEBUG oslo_vmware.api [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1150.460298] env[62952]: value = "task-1367729" [ 1150.460298] env[62952]: _type = "Task" [ 1150.460298] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.467905] env[62952]: DEBUG oslo_vmware.api [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367729, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.471569] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.715218] env[62952]: DEBUG nova.scheduler.client.report [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.973755] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.974193] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1150.974557] env[62952]: DEBUG oslo_vmware.api [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367729, 'name': ReconfigVM_Task, 'duration_secs': 0.145431} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.974888] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.975330] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290988', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'name': 'volume-16307e9b-ae51-423e-b715-36fca8845510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'serial': '16307e9b-ae51-423e-b715-36fca8845510'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1150.977289] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.977622] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.978189] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.978476] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.978759] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.978998] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1150.979283] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.220566] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.242972] env[62952]: INFO nova.scheduler.client.report [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance 2fa96864-5440-454c-a743-fdf4e70a0e2e [ 1151.361064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.361064] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.361564] env[62952]: INFO nova.compute.manager [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Attaching volume ecd5ca41-746f-4f21-b85b-ae74dec11fc6 to /dev/sdb [ 1151.392417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8361069d-3c68-4d36-af93-dd519c36dbf4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.399679] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6735f7-18ad-47ba-9ee9-a0a92d8b6d10 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.413419] env[62952]: DEBUG nova.virt.block_device [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating existing volume attachment record: 62139ec9-b271-43bd-a2bb-1cc2483d9d83 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1151.483797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.483797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.483797] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.483797] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1151.484396] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabef1b1-ef54-4af2-9293-673f732efa7d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.493017] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae532f2-abe7-4cc9-a152-36999ec087d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.505637] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905b025a-7604-4770-ad06-1b973c53c321 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.514016] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e8839c-3928-437c-a8f1-1ff76da96ab5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.544053] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179988MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1151.544208] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.544416] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.752203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74b5a640-6995-4b7b-8939-f8aa975880fd tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "2fa96864-5440-454c-a743-fdf4e70a0e2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.673s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.011887] env[62952]: DEBUG nova.objects.instance [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 287250a7-3354-49b7-9194-d88ae51afdaf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.516603] env[62952]: DEBUG oslo_concurrency.lockutils [None req-55629ec8-e9d3-44b9-8794-ece0d26759d3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.572415] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.572415] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 532a2bcd-5d6a-4fa9-abc9-a048cc915fda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.572591] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 287250a7-3354-49b7-9194-d88ae51afdaf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.572749] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance ff8301b2-c5cb-4e46-8f5c-98577dfd4835 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.572810] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.572887] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 26fdfac2-4ea1-4822-b38c-d8f7388436be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.573082] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1152.573228] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1152.654792] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038666e7-1689-44d8-8a47-fc99fb238c02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.662310] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba35788-c367-4fe1-b6a6-0d38344cdc30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.692177] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5288c39-0095-4ba6-9711-91914cf46c4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.699189] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ebcb61a-bc45-4855-81cd-e6a89ae361aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.711711] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.819203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.819484] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.819810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.820086] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.820226] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.822458] env[62952]: INFO nova.compute.manager [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Terminating instance [ 1152.824209] env[62952]: DEBUG nova.compute.manager [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1152.824406] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1152.825261] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200f280b-f1b0-44fe-8cf2-f0bb59c96fff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.833279] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1152.833512] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-101a49f5-a468-404d-8f91-61d77eb15552 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.839155] env[62952]: DEBUG oslo_vmware.api [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1152.839155] env[62952]: value = "task-1367735" [ 1152.839155] env[62952]: _type = "Task" [ 1152.839155] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.846401] env[62952]: DEBUG oslo_vmware.api [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.214648] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1153.340365] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.340682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.352263] env[62952]: DEBUG oslo_vmware.api [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367735, 'name': PowerOffVM_Task, 'duration_secs': 0.256775} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.353148] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1153.353332] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1153.353579] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8487374-5098-4670-9b73-d88b4ddda804 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.413566] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1153.413783] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1153.413969] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleting the datastore file [datastore2] 532a2bcd-5d6a-4fa9-abc9-a048cc915fda {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.414245] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f43ebd7-320e-4900-ab13-901958bd9c93 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.421414] env[62952]: DEBUG oslo_vmware.api [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for the task: (returnval){ [ 1153.421414] env[62952]: value = "task-1367737" [ 1153.421414] env[62952]: _type = "Task" [ 1153.421414] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.428601] env[62952]: DEBUG oslo_vmware.api [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.719393] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1153.719714] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.175s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.720026] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.720208] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Cleaning up deleted instances {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1153.843719] env[62952]: DEBUG nova.compute.utils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1153.930286] env[62952]: DEBUG oslo_vmware.api [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Task: {'id': task-1367737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14378} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.930525] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.930711] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1153.930890] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1153.931090] env[62952]: INFO nova.compute.manager [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1153.931336] env[62952]: DEBUG oslo.service.loopingcall [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.931529] env[62952]: DEBUG nova.compute.manager [-] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1153.931632] env[62952]: DEBUG nova.network.neutron [-] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1154.202938] env[62952]: DEBUG nova.compute.manager [req-0a6f27b3-af13-47cb-995a-e1fe82158543 req-9fde03c8-cd9e-48de-9a06-225965cb2a25 service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Received event network-vif-deleted-be66392d-0f28-4019-9db6-f989ba14dbe6 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.203261] env[62952]: INFO nova.compute.manager [req-0a6f27b3-af13-47cb-995a-e1fe82158543 req-9fde03c8-cd9e-48de-9a06-225965cb2a25 service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Neutron deleted interface be66392d-0f28-4019-9db6-f989ba14dbe6; detaching it from the instance and deleting it from the info cache [ 1154.203510] env[62952]: DEBUG nova.network.neutron [req-0a6f27b3-af13-47cb-995a-e1fe82158543 req-9fde03c8-cd9e-48de-9a06-225965cb2a25 service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.232820] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] There are 43 instances to clean {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1154.233116] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2fa96864-5440-454c-a743-fdf4e70a0e2e] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1154.346354] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.679345] env[62952]: DEBUG nova.network.neutron [-] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.706080] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bf2682d-7280-4b5f-ae94-3a982ac4a23b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.715304] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065a2b5f-06ef-4542-8c73-97e442d4f162 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.740606] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9036f863-9d77-491b-8aa9-978d0c400708] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1154.742307] env[62952]: DEBUG nova.compute.manager [req-0a6f27b3-af13-47cb-995a-e1fe82158543 req-9fde03c8-cd9e-48de-9a06-225965cb2a25 service nova] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Detach interface failed, port_id=be66392d-0f28-4019-9db6-f989ba14dbe6, reason: Instance 532a2bcd-5d6a-4fa9-abc9-a048cc915fda could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1155.182495] env[62952]: INFO nova.compute.manager [-] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Took 1.25 seconds to deallocate network for instance. [ 1155.244086] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: f302233b-2e1f-4c71-b000-0c9eb6530730] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1155.410849] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.411151] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.411396] env[62952]: INFO nova.compute.manager [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Attaching volume 75a9566d-028d-49c2-94e4-1ef05f47da42 to /dev/sdc [ 1155.440270] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4877d53-5ed3-49e9-95e1-f69d8a3ddbd4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.447274] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee56f26c-ec0c-42ab-923b-40ee2aa73fc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.459495] env[62952]: DEBUG nova.virt.block_device [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updating existing volume attachment record: 58e85358-6338-4105-b15b-8a61974f51b6 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1155.688915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.689312] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.689736] env[62952]: DEBUG nova.objects.instance [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lazy-loading 'resources' on Instance uuid 532a2bcd-5d6a-4fa9-abc9-a048cc915fda {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.746809] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 9c18c8c7-bc97-4a84-a86d-e23f8cce20ed] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1155.961030] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1155.961340] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290991', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'name': 'volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835', 'attached_at': '', 'detached_at': '', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'serial': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1155.962278] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666c8b24-3633-493d-a6f2-c2e23b5045aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.978384] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b660308-55ab-4fd9-b67d-1afdd8207f8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.002444] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6/volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.002692] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4490c6bc-f8cc-4bd1-98d2-b859a9f8e84e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.019955] env[62952]: DEBUG oslo_vmware.api [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1156.019955] env[62952]: value = "task-1367741" [ 1156.019955] env[62952]: _type = "Task" [ 1156.019955] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.027479] env[62952]: DEBUG oslo_vmware.api [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.250378] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 44f0943d-7267-4472-8fe2-445a0727654a] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1156.283026] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c5b929-4d3a-4f1d-ac37-383b6129c690 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.290961] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327702d9-33dc-4981-9a7f-58c8657df2dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.321861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004b01ae-bcdf-43d7-bf09-f3aa0db1973d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.329175] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa49718-236d-42e4-b465-089017bfee8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.342262] env[62952]: DEBUG nova.compute.provider_tree [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.529723] env[62952]: DEBUG oslo_vmware.api [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367741, 'name': ReconfigVM_Task, 'duration_secs': 0.476402} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.529946] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to attach disk [datastore1] volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6/volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.534614] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1237bbc-f90b-4eb7-b2a4-648118928fc1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.548125] env[62952]: DEBUG oslo_vmware.api [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1156.548125] env[62952]: value = "task-1367742" [ 1156.548125] env[62952]: _type = "Task" [ 1156.548125] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.555514] env[62952]: DEBUG oslo_vmware.api [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.754015] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 39eef6be-fe54-418b-b88d-104d082c3ca7] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1156.845824] env[62952]: DEBUG nova.scheduler.client.report [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.057662] env[62952]: DEBUG oslo_vmware.api [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367742, 'name': ReconfigVM_Task, 'duration_secs': 0.125745} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.057955] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290991', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'name': 'volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835', 'attached_at': '', 'detached_at': '', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'serial': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1157.257642] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: ea621ff2-e54e-4403-9e6d-dea84c7e8ad2] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1157.350689] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.370761] env[62952]: INFO nova.scheduler.client.report [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Deleted allocations for instance 532a2bcd-5d6a-4fa9-abc9-a048cc915fda [ 1157.761236] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: b9c57a92-73bc-4ee5-8431-2fe51a695383] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1157.878104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d5811906-be92-4b8b-a628-99be86942c43 tempest-ServersTestJSON-869811831 tempest-ServersTestJSON-869811831-project-member] Lock "532a2bcd-5d6a-4fa9-abc9-a048cc915fda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.058s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.091601] env[62952]: DEBUG nova.objects.instance [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'flavor' on Instance uuid ff8301b2-c5cb-4e46-8f5c-98577dfd4835 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.264696] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 12365f25-7d85-48dc-b2b7-9c8d0695e4a9] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1158.595974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e32e570a-3bf2-4e38-8495-d32f90559c65 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.769375] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: faa3b525-bd20-4e54-ba4c-60d65bf601aa] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1159.272393] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: f5e5c5c1-e2af-40e0-a957-67b0fd22718d] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1159.453124] env[62952]: DEBUG nova.compute.manager [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1159.775484] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2986a508-3f0b-4bad-b785-9e9f3ca7e39c] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1159.975670] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.975927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.001914] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1160.002179] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290992', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'name': 'volume-75a9566d-028d-49c2-94e4-1ef05f47da42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'serial': '75a9566d-028d-49c2-94e4-1ef05f47da42'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1160.003232] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03041bc-0097-4354-b487-af9d125a2611 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.019475] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6588a8c-f961-484c-a3c6-dfe50a424f11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.047026] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] volume-75a9566d-028d-49c2-94e4-1ef05f47da42/volume-75a9566d-028d-49c2-94e4-1ef05f47da42.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1160.047315] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-804858fc-c8a9-44e9-8e02-7cba919af01f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.064362] env[62952]: DEBUG oslo_vmware.api [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1160.064362] env[62952]: value = "task-1367745" [ 1160.064362] env[62952]: _type = "Task" [ 1160.064362] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.073815] env[62952]: DEBUG oslo_vmware.api [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367745, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.278267] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: e4fd9b9a-fc39-443c-9317-8dd286fdd2b0] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1160.481510] env[62952]: INFO nova.compute.claims [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1160.574887] env[62952]: DEBUG oslo_vmware.api [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367745, 'name': ReconfigVM_Task, 'duration_secs': 0.410307} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.575199] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfigured VM instance instance-00000064 to attach disk [datastore2] volume-75a9566d-028d-49c2-94e4-1ef05f47da42/volume-75a9566d-028d-49c2-94e4-1ef05f47da42.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.579981] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33fe80eb-3136-4303-a644-dce8f2dd6fc6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.597068] env[62952]: DEBUG oslo_vmware.api [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1160.597068] env[62952]: value = "task-1367746" [ 1160.597068] env[62952]: _type = "Task" [ 1160.597068] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.605576] env[62952]: DEBUG oslo_vmware.api [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367746, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.781526] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 1df00ddb-64fd-4ad3-be02-072a5aaad6f2] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1160.989568] env[62952]: INFO nova.compute.resource_tracker [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating resource usage from migration de73d5a1-7409-4413-b8be-1f26c2c2ff67 [ 1161.081172] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd8d3e5-dd38-427d-a9b4-fb64fe040246 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.088860] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dcd349-d14b-43d7-b273-c6bd4adc5ccc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.129029] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b64d50-d157-43a5-b30a-758f2575b552 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.137398] env[62952]: DEBUG oslo_vmware.api [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367746, 'name': ReconfigVM_Task, 'duration_secs': 0.136048} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.138984] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290992', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'name': 'volume-75a9566d-028d-49c2-94e4-1ef05f47da42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'serial': '75a9566d-028d-49c2-94e4-1ef05f47da42'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1161.141634] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e056fb6-b686-466c-9998-008fbf504eaa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.159606] env[62952]: DEBUG nova.compute.provider_tree [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.284774] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: e0384d0d-6738-47f7-b14d-994cae2d1fe4] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1161.668531] env[62952]: DEBUG nova.scheduler.client.report [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1161.788427] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 75d640ed-c41a-4761-8867-191d8b3e1f79] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1162.173973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.198s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.174222] env[62952]: INFO nova.compute.manager [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Migrating [ 1162.192226] env[62952]: DEBUG nova.objects.instance [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 287250a7-3354-49b7-9194-d88ae51afdaf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.292129] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: ba2dfb85-0ffa-4883-abf1-482441c5afb8] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1162.387032] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.387247] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.691526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.691602] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.691793] env[62952]: DEBUG nova.network.neutron [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1162.698921] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3664f113-5dfd-44c9-a5e1-6fd236a9bbc3 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.288s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.797972] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 4c4845e6-79ca-4a2a-a636-842d6919e70f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1162.891679] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1163.067995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.068135] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.301463] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 1e02f16d-b86e-4533-928b-a672ce69788f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1163.411427] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.411862] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.413242] env[62952]: INFO nova.compute.claims [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1163.416460] env[62952]: DEBUG nova.network.neutron [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.558248] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.558430] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.570857] env[62952]: INFO nova.compute.manager [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Detaching volume 16307e9b-ae51-423e-b715-36fca8845510 [ 1163.609752] env[62952]: INFO nova.virt.block_device [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Attempting to driver detach volume 16307e9b-ae51-423e-b715-36fca8845510 from mountpoint /dev/sdb [ 1163.610010] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1163.610223] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290988', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'name': 'volume-16307e9b-ae51-423e-b715-36fca8845510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'serial': '16307e9b-ae51-423e-b715-36fca8845510'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1163.611134] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de1e466-d7db-4749-b2d3-bf7675624d99 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.634675] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8194744d-ce88-46e5-b945-176d1d2e5283 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.643139] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0188b4cc-4dc5-4e4e-bd13-de11234602b1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.665196] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd01c95-870f-4b5b-bd2f-cc69c5b1e937 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.679135] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] The volume has not been displaced from its original location: [datastore1] volume-16307e9b-ae51-423e-b715-36fca8845510/volume-16307e9b-ae51-423e-b715-36fca8845510.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1163.684419] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1163.684603] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bff30da6-df34-4446-8396-745a741e2e75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.702016] env[62952]: DEBUG oslo_vmware.api [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1163.702016] env[62952]: value = "task-1367747" [ 1163.702016] env[62952]: _type = "Task" [ 1163.702016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.710438] env[62952]: DEBUG oslo_vmware.api [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.805333] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: da749732-72d8-4180-aedf-73a17fa9dea6] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1163.922196] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.061474] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1164.211669] env[62952]: DEBUG oslo_vmware.api [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367747, 'name': ReconfigVM_Task, 'duration_secs': 0.225868} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.211946] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1164.216521] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b36849a0-6d27-4b6b-a806-efcea582f4c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.230831] env[62952]: DEBUG oslo_vmware.api [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1164.230831] env[62952]: value = "task-1367748" [ 1164.230831] env[62952]: _type = "Task" [ 1164.230831] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.238323] env[62952]: DEBUG oslo_vmware.api [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367748, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.309113] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 244f7b73-fda4-4477-b04d-c4f5cedd2d98] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1164.538181] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264abcd3-4fed-49f4-8efc-86c2b5d58ac8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.545520] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049e4b65-0df9-4e99-aa52-2b5c01912c8f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.578620] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d448cbfc-1400-4b67-89fb-836c20e21d26 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.587339] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723a802f-c70c-4232-af17-5716b21e0adb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.592555] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.600914] env[62952]: DEBUG nova.compute.provider_tree [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.742507] env[62952]: DEBUG oslo_vmware.api [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367748, 'name': ReconfigVM_Task, 'duration_secs': 0.13969} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.742859] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290988', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'name': 'volume-16307e9b-ae51-423e-b715-36fca8845510', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '16307e9b-ae51-423e-b715-36fca8845510', 'serial': '16307e9b-ae51-423e-b715-36fca8845510'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1164.812509] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 303c9fe7-0d89-4a1a-adcc-6f787486df61] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1165.105458] env[62952]: DEBUG nova.scheduler.client.report [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1165.287690] env[62952]: DEBUG nova.objects.instance [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 287250a7-3354-49b7-9194-d88ae51afdaf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.315767] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 97995e38-b27e-478a-8553-eb1c844bb0a6] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1165.438901] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04742713-d990-4155-bbb0-be673daa4ed2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.462065] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1165.608495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.608911] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1165.611533] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.019s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.613356] env[62952]: INFO nova.compute.claims [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1165.818648] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 82401449-347e-4917-a307-d5d7bc048f7c] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1165.968407] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.968732] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ef255b5-f64d-4a58-8ac6-e2761ba27f73 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.976171] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1165.976171] env[62952]: value = "task-1367749" [ 1165.976171] env[62952]: _type = "Task" [ 1165.976171] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.984157] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367749, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.118414] env[62952]: DEBUG nova.compute.utils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1166.122201] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1166.122385] env[62952]: DEBUG nova.network.neutron [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1166.173800] env[62952]: DEBUG nova.policy [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '913eacf6a94845b189e0bd5df408138a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8c2a296e8fe4805b9bb761f91a407f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1166.295418] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77c230fe-0842-47be-80be-90c119ef13a7 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.227s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.322653] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 3bd9c356-42d6-450f-9069-b066952d9524] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1166.406356] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.407046] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.485885] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367749, 'name': PowerOffVM_Task, 'duration_secs': 0.198546} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.486671] env[62952]: DEBUG nova.network.neutron [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Successfully created port: de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1166.488450] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1166.488656] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1166.623759] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1166.730000] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff59761c-0f9e-4acf-84bf-492e12766b65 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.737508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4623bebf-551e-4a7e-88be-f5befbd99430 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.765640] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67b0348-1198-4269-9088-d46cfeb58ccf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.772264] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd3bb2a-3617-42ba-b17f-72a739527bdb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.784557] env[62952]: DEBUG nova.compute.provider_tree [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.826268] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 8b2e4601-b981-48d1-84e6-21da66d5ea5f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1166.909388] env[62952]: INFO nova.compute.manager [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Detaching volume 75a9566d-028d-49c2-94e4-1ef05f47da42 [ 1166.945721] env[62952]: INFO nova.virt.block_device [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Attempting to driver detach volume 75a9566d-028d-49c2-94e4-1ef05f47da42 from mountpoint /dev/sdc [ 1166.946015] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1166.946220] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290992', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'name': 'volume-75a9566d-028d-49c2-94e4-1ef05f47da42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'serial': '75a9566d-028d-49c2-94e4-1ef05f47da42'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1166.947104] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139c580e-bda3-4037-b29a-04093f902c37 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.967717] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1ed81e-29c3-4fbc-9568-06a602461f67 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.975751] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d715ebb8-9713-4d6a-9ae4-178224e4c38a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.996550] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1166.996781] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1166.996946] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1166.997165] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1166.997310] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1166.997464] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1166.997669] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1166.997829] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1166.998436] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1166.998436] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1166.998436] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1167.003466] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73c23575-2ec2-410c-8113-b30985439c03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.013508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5324c964-03e0-479c-833e-bf27d2ec0802 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.029437] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] The volume has not been displaced from its original location: [datastore2] volume-75a9566d-028d-49c2-94e4-1ef05f47da42/volume-75a9566d-028d-49c2-94e4-1ef05f47da42.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1167.034656] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfiguring VM instance instance-00000064 to detach disk 2002 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1167.036256] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-896d7bbd-a2ee-4630-8a48-0d065e0e7c23 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.051103] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1167.051103] env[62952]: value = "task-1367750" [ 1167.051103] env[62952]: _type = "Task" [ 1167.051103] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.056613] env[62952]: DEBUG oslo_vmware.api [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1167.056613] env[62952]: value = "task-1367751" [ 1167.056613] env[62952]: _type = "Task" [ 1167.056613] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.059923] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.067330] env[62952]: DEBUG oslo_vmware.api [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367751, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.133997] env[62952]: INFO nova.virt.block_device [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Booting with volume 188d94f2-707a-48de-b891-ba671ade4185 at /dev/sda [ 1167.167038] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfe97a65-07c6-42a6-9799-0ba0f198ea36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.175865] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4671293d-c115-47f6-ae5b-68df8d63947a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.202807] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-284aed9d-204e-4304-82a1-7e586ab23202 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.211145] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dad57d1-b882-4d2a-9a2b-650d28d7b99a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.243841] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d882156-bcb2-4049-8b72-0bf28c637cd7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.253903] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546056f2-f97f-4efb-826c-2ac32034345a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.270721] env[62952]: DEBUG nova.virt.block_device [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating existing volume attachment record: 0c62097d-e896-40b3-98b7-b738157dd307 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1167.288152] env[62952]: DEBUG nova.scheduler.client.report [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.329028] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: ff189c9e-9e7c-4217-9c65-0f821393870a] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1167.563073] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367750, 'name': ReconfigVM_Task, 'duration_secs': 0.206424} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.566057] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1167.574287] env[62952]: DEBUG oslo_vmware.api [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367751, 'name': ReconfigVM_Task, 'duration_secs': 0.228757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.574537] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Reconfigured VM instance instance-00000064 to detach disk 2002 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1167.579133] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ac3d389-67bf-448c-bc2d-b2ced9db8df6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.594303] env[62952]: DEBUG oslo_vmware.api [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1167.594303] env[62952]: value = "task-1367752" [ 1167.594303] env[62952]: _type = "Task" [ 1167.594303] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.602819] env[62952]: DEBUG oslo_vmware.api [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367752, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.795053] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.795896] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1167.833223] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 867ffed2-50ec-45d8-b64e-989cf8d1b0e2] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1167.864110] env[62952]: DEBUG nova.compute.manager [req-07d1e368-8b1d-4d67-be94-378d9ddb9b07 req-2ea9c9be-d836-4dd5-8514-398d060913f1 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Received event network-vif-plugged-de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1167.864276] env[62952]: DEBUG oslo_concurrency.lockutils [req-07d1e368-8b1d-4d67-be94-378d9ddb9b07 req-2ea9c9be-d836-4dd5-8514-398d060913f1 service nova] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.864491] env[62952]: DEBUG oslo_concurrency.lockutils [req-07d1e368-8b1d-4d67-be94-378d9ddb9b07 req-2ea9c9be-d836-4dd5-8514-398d060913f1 service nova] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.864666] env[62952]: DEBUG oslo_concurrency.lockutils [req-07d1e368-8b1d-4d67-be94-378d9ddb9b07 req-2ea9c9be-d836-4dd5-8514-398d060913f1 service nova] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.864833] env[62952]: DEBUG nova.compute.manager [req-07d1e368-8b1d-4d67-be94-378d9ddb9b07 req-2ea9c9be-d836-4dd5-8514-398d060913f1 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] No waiting events found dispatching network-vif-plugged-de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1167.865009] env[62952]: WARNING nova.compute.manager [req-07d1e368-8b1d-4d67-be94-378d9ddb9b07 req-2ea9c9be-d836-4dd5-8514-398d060913f1 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Received unexpected event network-vif-plugged-de8307b9-6263-4138-befe-40f603f1bea4 for instance with vm_state building and task_state block_device_mapping. [ 1167.952506] env[62952]: DEBUG nova.network.neutron [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Successfully updated port: de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1168.072974] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1168.073289] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1168.073480] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1168.073725] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1168.073931] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1168.074153] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1168.074412] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1168.074639] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1168.074872] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1168.075097] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1168.075351] env[62952]: DEBUG nova.virt.hardware [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1168.081344] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1168.081637] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5f5717e-87bd-4c96-becc-0e6467cef231 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.104039] env[62952]: DEBUG oslo_vmware.api [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367752, 'name': ReconfigVM_Task, 'duration_secs': 0.128819} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.105333] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290992', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'name': 'volume-75a9566d-028d-49c2-94e4-1ef05f47da42', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '287250a7-3354-49b7-9194-d88ae51afdaf', 'attached_at': '', 'detached_at': '', 'volume_id': '75a9566d-028d-49c2-94e4-1ef05f47da42', 'serial': '75a9566d-028d-49c2-94e4-1ef05f47da42'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1168.107584] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1168.107584] env[62952]: value = "task-1367753" [ 1168.107584] env[62952]: _type = "Task" [ 1168.107584] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.115515] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367753, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.303451] env[62952]: DEBUG nova.compute.utils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1168.304896] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1168.305080] env[62952]: DEBUG nova.network.neutron [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1168.337907] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: a88447f6-6a5a-490d-bf44-11c9237fa5cb] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1168.349904] env[62952]: DEBUG nova.policy [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77db11499b634ddf847014f4f4c169f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2c2a97031df4b75b835a96b41793c74', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1168.455090] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.455528] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.455742] env[62952]: DEBUG nova.network.neutron [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1168.618646] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367753, 'name': ReconfigVM_Task, 'duration_secs': 0.187457} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.619028] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1168.619756] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694786ed-f12e-431a-8f8f-305894e4b4a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.646058] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1168.646811] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-778004ea-a25e-4f2d-8cb3-4c32873248b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.660339] env[62952]: DEBUG nova.objects.instance [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'flavor' on Instance uuid 287250a7-3354-49b7-9194-d88ae51afdaf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.667612] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1168.667612] env[62952]: value = "task-1367754" [ 1168.667612] env[62952]: _type = "Task" [ 1168.667612] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.668402] env[62952]: DEBUG nova.network.neutron [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Successfully created port: 551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1168.678531] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367754, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.808655] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1168.843739] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: a3c01d4d-9c66-4441-95e8-87998782cc02] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1168.986879] env[62952]: DEBUG nova.network.neutron [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1169.117514] env[62952]: DEBUG nova.network.neutron [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.180382] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367754, 'name': ReconfigVM_Task, 'duration_secs': 0.317436} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.180660] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to attach disk [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.180941] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1169.345522] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 88cd1c0d-a0e6-4cb8-8da6-c53d044c7e6b] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1169.357487] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1169.358174] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1169.358507] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1169.358797] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1169.359080] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1169.359278] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1169.359474] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1169.359729] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1169.359980] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1169.360193] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1169.360361] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1169.360719] env[62952]: DEBUG nova.virt.hardware [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1169.361497] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd703b6-f2b5-44d2-86aa-a6107547edd7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.370932] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5171d746-ce69-46d2-9209-ae42657cad85 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.620689] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.621096] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Instance network_info: |[{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1169.621539] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:94:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de8307b9-6263-4138-befe-40f603f1bea4', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1169.629145] env[62952]: DEBUG oslo.service.loopingcall [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1169.629397] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1169.629632] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f26a4bc6-d664-4207-8aaa-274e3d30d205 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.649933] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1169.649933] env[62952]: value = "task-1367755" [ 1169.649933] env[62952]: _type = "Task" [ 1169.649933] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.657446] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367755, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.667026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-8aadef70-f837-4d97-9c46-c6c748d22627 tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.260s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.686994] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8262cc-1103-4e69-8906-2cc2c6001a14 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.708724] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34717490-f72f-4cb0-ad91-e3f6ae4a0945 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.729038] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1169.817904] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1169.843413] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1169.843674] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1169.843837] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1169.844087] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1169.844182] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1169.844336] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1169.844542] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1169.844706] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1169.844878] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1169.845054] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1169.845240] env[62952]: DEBUG nova.virt.hardware [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1169.846109] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713b4ac7-87af-4a48-8b73-b1dfa57a6752 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.848933] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 94fd5287-9a8e-45e4-99e8-0bcc861f889c] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1169.856978] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a21cff-ad30-4cd3-aae8-3dda04a262a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.889437] env[62952]: DEBUG nova.compute.manager [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Received event network-changed-de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1169.889659] env[62952]: DEBUG nova.compute.manager [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Refreshing instance network info cache due to event network-changed-de8307b9-6263-4138-befe-40f603f1bea4. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1169.890030] env[62952]: DEBUG oslo_concurrency.lockutils [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] Acquiring lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1169.890193] env[62952]: DEBUG oslo_concurrency.lockutils [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] Acquired lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.890368] env[62952]: DEBUG nova.network.neutron [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Refreshing network info cache for port de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1170.160228] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367755, 'name': CreateVM_Task, 'duration_secs': 0.346271} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.160398] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1170.161100] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'attachment_id': '0c62097d-e896-40b3-98b7-b738157dd307', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290989', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'name': 'volume-188d94f2-707a-48de-b891-ba671ade4185', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6606e84c-36cd-4519-beff-096a8f1bc09e', 'attached_at': '', 'detached_at': '', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'serial': '188d94f2-707a-48de-b891-ba671ade4185'}, 'guest_format': None, 'device_type': None, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62952) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1170.161329] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Root volume attach. Driver type: vmdk {{(pid=62952) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1170.162087] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f805d195-4a4f-4c64-ab7a-c09cef02a1a6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.169597] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e64e177-ec9d-45b8-b98f-b7c88fea5ff8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.175230] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee1fb8f-a92f-4603-9f83-bf5d2d35e76b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.178983] env[62952]: DEBUG nova.network.neutron [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Successfully updated port: 551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1170.183351] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e9bf27f6-5775-4892-99c5-9277eebcf51d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.188973] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1170.188973] env[62952]: value = "task-1367756" [ 1170.188973] env[62952]: _type = "Task" [ 1170.188973] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.197607] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.354139] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 93951b09-9d7c-42da-83de-0db3980d0d5d] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1170.654041] env[62952]: DEBUG nova.network.neutron [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updated VIF entry in instance network info cache for port de8307b9-6263-4138-befe-40f603f1bea4. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1170.654041] env[62952]: DEBUG nova.network.neutron [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.683270] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.683597] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.683597] env[62952]: DEBUG nova.network.neutron [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1170.700114] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 42%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.799665] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.800045] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.800815] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "287250a7-3354-49b7-9194-d88ae51afdaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.801071] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.801337] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.804197] env[62952]: INFO nova.compute.manager [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Terminating instance [ 1170.806449] env[62952]: DEBUG nova.compute.manager [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1170.806732] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.807790] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db330f9-c767-4782-8acf-1692b6818769 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.817063] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1170.817384] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40d72ef6-a771-4165-a949-87f7c3f5f520 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.825143] env[62952]: DEBUG oslo_vmware.api [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1170.825143] env[62952]: value = "task-1367757" [ 1170.825143] env[62952]: _type = "Task" [ 1170.825143] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.835120] env[62952]: DEBUG oslo_vmware.api [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.857780] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 7621d6a6-27aa-45d3-80fa-65e957519a1f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1171.156857] env[62952]: DEBUG oslo_concurrency.lockutils [req-62a88a83-0f15-4ec6-b79e-8b5d94754af9 req-bada6bc9-25ad-453e-991c-bf7cef07c6a6 service nova] Releasing lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.201590] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 56%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.218804] env[62952]: DEBUG nova.network.neutron [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1171.335046] env[62952]: DEBUG oslo_vmware.api [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367757, 'name': PowerOffVM_Task, 'duration_secs': 0.236648} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.335334] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1171.335509] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1171.335813] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23fcb5b8-930c-47cb-962a-367bbc13a09a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.361516] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 4983f354-77dc-4e84-969b-4c74441fa568] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1171.377269] env[62952]: DEBUG nova.network.neutron [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Port 612c5fea-4183-448f-9326-6ed1888b591d binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1171.379802] env[62952]: DEBUG nova.network.neutron [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.403527] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1171.403748] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1171.403942] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Deleting the datastore file [datastore2] 287250a7-3354-49b7-9194-d88ae51afdaf {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1171.404229] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a81a6f46-03ec-4207-86f5-950923da058b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.412356] env[62952]: DEBUG oslo_vmware.api [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for the task: (returnval){ [ 1171.412356] env[62952]: value = "task-1367759" [ 1171.412356] env[62952]: _type = "Task" [ 1171.412356] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.421426] env[62952]: DEBUG oslo_vmware.api [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.699735] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 71%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.868847] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 34f5995f-692c-4f0d-8b15-0d388df7e34e] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1171.886203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.886525] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance network_info: |[{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1171.887191] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e2:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd22cb4ec-277f-41ee-8aba-b3d54442b93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '551a3758-07da-4c7a-acda-583eb14f995f', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1171.896036] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating folder: Project (a2c2a97031df4b75b835a96b41793c74). Parent ref: group-v290852. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1171.896036] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a48b20f9-b299-47ef-b8ed-768abbf405de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.907556] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created folder: Project (a2c2a97031df4b75b835a96b41793c74) in parent group-v290852. [ 1171.907785] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating folder: Instances. Parent ref: group-v290994. {{(pid=62952) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1171.908048] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2fb09e7f-d004-4779-b7ee-2e2a78a466d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.919162] env[62952]: DEBUG nova.compute.manager [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-vif-plugged-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.919394] env[62952]: DEBUG oslo_concurrency.lockutils [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.919640] env[62952]: DEBUG oslo_concurrency.lockutils [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.919858] env[62952]: DEBUG oslo_concurrency.lockutils [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.920048] env[62952]: DEBUG nova.compute.manager [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] No waiting events found dispatching network-vif-plugged-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1171.920219] env[62952]: WARNING nova.compute.manager [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received unexpected event network-vif-plugged-551a3758-07da-4c7a-acda-583eb14f995f for instance with vm_state building and task_state spawning. [ 1171.920382] env[62952]: DEBUG nova.compute.manager [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-changed-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.920538] env[62952]: DEBUG nova.compute.manager [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing instance network info cache due to event network-changed-551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1171.920726] env[62952]: DEBUG oslo_concurrency.lockutils [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.920868] env[62952]: DEBUG oslo_concurrency.lockutils [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.921045] env[62952]: DEBUG nova.network.neutron [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1171.924157] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created folder: Instances in parent group-v290994. [ 1171.924419] env[62952]: DEBUG oslo.service.loopingcall [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.924906] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1171.928170] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1db5662d-7f91-49e9-b7da-09d461025ed6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.943101] env[62952]: DEBUG oslo_vmware.api [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Task: {'id': task-1367759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154786} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.944414] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.944621] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1171.944805] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.944985] env[62952]: INFO nova.compute.manager [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1171.945251] env[62952]: DEBUG oslo.service.loopingcall [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.946115] env[62952]: DEBUG nova.compute.manager [-] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.946220] env[62952]: DEBUG nova.network.neutron [-] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1171.950098] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1171.950098] env[62952]: value = "task-1367762" [ 1171.950098] env[62952]: _type = "Task" [ 1171.950098] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.958624] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367762, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.200510] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 86%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.374297] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 5914a35a-0934-4f8f-81e8-d91bc690a9cb] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1172.410097] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.410426] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.410657] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.464456] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367762, 'name': CreateVM_Task, 'duration_secs': 0.379066} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.464616] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1172.465306] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.465496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.465832] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1172.466069] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf17a38e-d0d6-4361-a000-cb0c7a0ebaeb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.470847] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1172.470847] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5239fa20-3423-34fa-1a5e-c72d69e371c3" [ 1172.470847] env[62952]: _type = "Task" [ 1172.470847] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.478796] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5239fa20-3423-34fa-1a5e-c72d69e371c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.623421] env[62952]: DEBUG nova.network.neutron [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updated VIF entry in instance network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1172.623827] env[62952]: DEBUG nova.network.neutron [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.700848] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 97%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.875622] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2f475040-28df-47e4-bfc4-4c843a57885f] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1172.914250] env[62952]: DEBUG nova.network.neutron [-] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.981197] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5239fa20-3423-34fa-1a5e-c72d69e371c3, 'name': SearchDatastore_Task, 'duration_secs': 0.009744} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.981535] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.981774] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1172.982046] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.982202] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.982381] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1172.982645] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4dac3d7-86ff-4451-846c-fd53c22b1172 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.991029] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1172.991029] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1172.991654] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-674099eb-af27-48ab-a573-03cef45b87f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.996695] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1172.996695] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52055b9d-e13f-a471-0f5d-c4e8cf546c07" [ 1172.996695] env[62952]: _type = "Task" [ 1172.996695] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.004462] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52055b9d-e13f-a471-0f5d-c4e8cf546c07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.129026] env[62952]: DEBUG oslo_concurrency.lockutils [req-6bf459ab-162e-4c75-8450-f32ba4e09827 req-03331cac-6409-4ffa-9d08-405a4b45c458 service nova] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.200854] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 97%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.378835] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 802fd770-38aa-4417-b25e-fd83510bceef] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1173.416048] env[62952]: INFO nova.compute.manager [-] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Took 1.47 seconds to deallocate network for instance. [ 1173.462907] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.463110] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.463508] env[62952]: DEBUG nova.network.neutron [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1173.507924] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52055b9d-e13f-a471-0f5d-c4e8cf546c07, 'name': SearchDatastore_Task, 'duration_secs': 0.008865} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.508853] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccd50c19-3a38-4cf8-90d0-67516731f1ee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.514220] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1173.514220] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b941da-63dc-b27e-b5a0-4565dca39dc7" [ 1173.514220] env[62952]: _type = "Task" [ 1173.514220] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.522200] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b941da-63dc-b27e-b5a0-4565dca39dc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.701774] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task} progress is 98%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.883040] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: a63ec2ac-4484-4360-962b-105a5338c1b8] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1173.923309] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.923594] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.923839] env[62952]: DEBUG nova.objects.instance [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lazy-loading 'resources' on Instance uuid 287250a7-3354-49b7-9194-d88ae51afdaf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.945107] env[62952]: DEBUG nova.compute.manager [req-7f666372-7af0-4a3b-94f9-f708a666168a req-2c177871-1de2-4c50-acbc-98f09feebbf1 service nova] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Received event network-vif-deleted-2595a3f5-136a-4d28-83ff-9bcc6d05154e {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1174.024183] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b941da-63dc-b27e-b5a0-4565dca39dc7, 'name': SearchDatastore_Task, 'duration_secs': 0.008316} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.024465] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.024753] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1174.026849] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e5078cf-02ee-4bf0-8f99-644203e94056 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.033312] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1174.033312] env[62952]: value = "task-1367763" [ 1174.033312] env[62952]: _type = "Task" [ 1174.033312] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.040458] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.177262] env[62952]: DEBUG nova.network.neutron [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.203816] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367756, 'name': RelocateVM_Task, 'duration_secs': 3.758161} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.204141] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1174.204355] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290989', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'name': 'volume-188d94f2-707a-48de-b891-ba671ade4185', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6606e84c-36cd-4519-beff-096a8f1bc09e', 'attached_at': '', 'detached_at': '', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'serial': '188d94f2-707a-48de-b891-ba671ade4185'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1174.205195] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7814650f-3a91-4a53-a870-20440356831a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.223625] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88118f1-5b68-4e64-95f3-404054d439f1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.247091] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-188d94f2-707a-48de-b891-ba671ade4185/volume-188d94f2-707a-48de-b891-ba671ade4185.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.247299] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc41b08c-82d3-4b30-a218-9eb6a3753a1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.268392] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1174.268392] env[62952]: value = "task-1367764" [ 1174.268392] env[62952]: _type = "Task" [ 1174.268392] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.278064] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367764, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.385796] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: d2faf1e0-cfd1-4d87-ba77-0af92dc16643] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1174.538199] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae60de1-16ab-467c-a43a-c268f421e1d1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.548386] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748e155c-9ad9-4609-83e0-f3a443a5dccb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.551366] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46137} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.551658] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1174.551921] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1174.552459] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04712f76-1a46-4cec-8563-826603c130f9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.580258] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164bd4b1-9a13-4b81-b97c-7fec64de0e45 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.583656] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1174.583656] env[62952]: value = "task-1367765" [ 1174.583656] env[62952]: _type = "Task" [ 1174.583656] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.589837] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a33f177-61ee-473a-a839-2c479d0411a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.596068] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367765, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.606321] env[62952]: DEBUG nova.compute.provider_tree [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.680459] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.778971] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367764, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.889826] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 351b1672-ddb3-4562-9965-d65a7d869e82] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1175.093386] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367765, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103872} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.093668] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1175.094449] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ceadde-e2a6-49fc-9367-6eecdb82f68d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.109857] env[62952]: DEBUG nova.scheduler.client.report [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1175.120911] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1175.121355] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b440e5c-4d87-46ab-81ea-6ed42ca09470 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.140922] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1175.140922] env[62952]: value = "task-1367766" [ 1175.140922] env[62952]: _type = "Task" [ 1175.140922] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.148796] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367766, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.190987] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5f021f-b4cf-44c5-89f2-907c0b8241ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.198470] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874ddab4-0c18-405c-b1b6-81b011f3e324 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.277992] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367764, 'name': ReconfigVM_Task, 'duration_secs': 0.779231} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.278344] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-188d94f2-707a-48de-b891-ba671ade4185/volume-188d94f2-707a-48de-b891-ba671ade4185.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.283206] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e857a41e-74ca-4069-8ef3-87adb238dd2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.297334] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1175.297334] env[62952]: value = "task-1367767" [ 1175.297334] env[62952]: _type = "Task" [ 1175.297334] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.305253] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367767, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.393655] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 34b6c13c-c95b-4cc2-b6eb-cc9374535aec] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1175.623203] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.646125] env[62952]: INFO nova.scheduler.client.report [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Deleted allocations for instance 287250a7-3354-49b7-9194-d88ae51afdaf [ 1175.652612] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367766, 'name': ReconfigVM_Task, 'duration_secs': 0.28007} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.653064] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.653656] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-483a4d5a-f108-42f6-b5ff-f38c9d34ef13 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.661343] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1175.661343] env[62952]: value = "task-1367768" [ 1175.661343] env[62952]: _type = "Task" [ 1175.661343] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.670048] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367768, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.806070] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367767, 'name': ReconfigVM_Task, 'duration_secs': 0.133003} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.806390] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290989', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'name': 'volume-188d94f2-707a-48de-b891-ba671ade4185', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6606e84c-36cd-4519-beff-096a8f1bc09e', 'attached_at': '', 'detached_at': '', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'serial': '188d94f2-707a-48de-b891-ba671ade4185'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1175.806928] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eee3d289-1b9e-4433-94df-7098ac9c69e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.813392] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1175.813392] env[62952]: value = "task-1367769" [ 1175.813392] env[62952]: _type = "Task" [ 1175.813392] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.821068] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367769, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.897181] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.897181] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Cleaning up deleted instances with incomplete migration {{(pid=62952) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1176.156170] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9249fb6c-0016-40fb-bbe1-ecacd48cf0bd tempest-AttachVolumeTestJSON-476661206 tempest-AttachVolumeTestJSON-476661206-project-member] Lock "287250a7-3354-49b7-9194-d88ae51afdaf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.356s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.172184] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367768, 'name': Rename_Task, 'duration_secs': 0.129872} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.172492] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1176.172749] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5475fca-5b0b-486b-b1ff-1ac98c9e9f49 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.179530] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1176.179530] env[62952]: value = "task-1367770" [ 1176.179530] env[62952]: _type = "Task" [ 1176.179530] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.188219] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.302072] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47577207-1fa6-4683-ac5e-de5919b6fb81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.327792] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981dc3c9-c923-43c1-9f5c-0743e0058d58 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.335944] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367769, 'name': Rename_Task, 'duration_secs': 0.129994} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.338778] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1176.339159] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1176.342676] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0f439f1-b932-44b5-bd96-4b3981e54ac4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.348489] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1176.348489] env[62952]: value = "task-1367772" [ 1176.348489] env[62952]: _type = "Task" [ 1176.348489] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.355644] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.400590] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1176.689768] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367770, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.846913] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1176.847375] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fc8a37b-e260-4486-8d8a-0547391cb8f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.859669] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367772, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.861081] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1176.861081] env[62952]: value = "task-1367773" [ 1176.861081] env[62952]: _type = "Task" [ 1176.861081] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.871146] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.190344] env[62952]: DEBUG oslo_vmware.api [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367770, 'name': PowerOnVM_Task, 'duration_secs': 0.685812} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.190689] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1177.190845] env[62952]: INFO nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1177.191040] env[62952]: DEBUG nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.191842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd474c5-81ca-4b5f-a3a0-60b90d5bde2d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.359523] env[62952]: DEBUG oslo_vmware.api [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367772, 'name': PowerOnVM_Task, 'duration_secs': 0.550264} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.359838] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1177.360059] env[62952]: INFO nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1177.360246] env[62952]: DEBUG nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.361059] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8605db9-20af-4832-8bc3-bbb45c3d39bb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.372960] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367773, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.708434] env[62952]: INFO nova.compute.manager [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Took 13.13 seconds to build instance. [ 1177.873200] env[62952]: DEBUG oslo_vmware.api [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367773, 'name': PowerOnVM_Task, 'duration_secs': 0.82917} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.873506] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1177.873792] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4016ea0f-9e97-4a12-81b3-036a98009bbc tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1177.883776] env[62952]: INFO nova.compute.manager [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Took 14.49 seconds to build instance. [ 1178.210593] env[62952]: DEBUG oslo_concurrency.lockutils [None req-0b773ab4-e62c-4d17-8307-99a52237a8c2 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.652s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.387693] env[62952]: DEBUG oslo_concurrency.lockutils [None req-76822fe8-0853-4b70-9f22-81408403dbf9 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.999s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.063249] env[62952]: DEBUG nova.compute.manager [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-changed-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1179.063249] env[62952]: DEBUG nova.compute.manager [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing instance network info cache due to event network-changed-551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1179.063893] env[62952]: DEBUG oslo_concurrency.lockutils [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.064364] env[62952]: DEBUG oslo_concurrency.lockutils [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.065080] env[62952]: DEBUG nova.network.neutron [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1179.094124] env[62952]: DEBUG nova.compute.manager [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Received event network-changed-f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1179.094124] env[62952]: DEBUG nova.compute.manager [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Refreshing instance network info cache due to event network-changed-f80edb6b-5326-4531-a02f-366484e828ce. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1179.094124] env[62952]: DEBUG oslo_concurrency.lockutils [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] Acquiring lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.094124] env[62952]: DEBUG oslo_concurrency.lockutils [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] Acquired lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.094277] env[62952]: DEBUG nova.network.neutron [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Refreshing network info cache for port f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1179.804929] env[62952]: DEBUG nova.network.neutron [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updated VIF entry in instance network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1179.805443] env[62952]: DEBUG nova.network.neutron [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.861434] env[62952]: DEBUG nova.network.neutron [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updated VIF entry in instance network info cache for port f80edb6b-5326-4531-a02f-366484e828ce. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1179.861825] env[62952]: DEBUG nova.network.neutron [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [{"id": "f80edb6b-5326-4531-a02f-366484e828ce", "address": "fa:16:3e:42:09:7a", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80edb6b-53", "ovs_interfaceid": "f80edb6b-5326-4531-a02f-366484e828ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.308961] env[62952]: DEBUG oslo_concurrency.lockutils [req-fee419a4-1b65-4239-9929-dcd6e0192dff req-3cba8b62-9e39-49e3-af82-c03624047bbc service nova] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.364875] env[62952]: DEBUG oslo_concurrency.lockutils [req-594ecf73-d31e-466c-8494-33373145f62c req-e0bf68c2-5eb4-4d10-86d6-a5e6e7a5cbee service nova] Releasing lock "refresh_cache-aef28168-98a7-4f65-80e7-731633339abf" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.445110] env[62952]: DEBUG nova.network.neutron [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Port 612c5fea-4183-448f-9326-6ed1888b591d binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1180.445408] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.445573] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.445835] env[62952]: DEBUG nova.network.neutron [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1180.572945] env[62952]: DEBUG nova.compute.manager [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Stashing vm_state: active {{(pid=62952) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1181.096741] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.097065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.121475] env[62952]: DEBUG nova.compute.manager [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Received event network-changed-de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1181.121677] env[62952]: DEBUG nova.compute.manager [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Refreshing instance network info cache due to event network-changed-de8307b9-6263-4138-befe-40f603f1bea4. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1181.121888] env[62952]: DEBUG oslo_concurrency.lockutils [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] Acquiring lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.122051] env[62952]: DEBUG oslo_concurrency.lockutils [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] Acquired lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.122225] env[62952]: DEBUG nova.network.neutron [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Refreshing network info cache for port de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1181.181340] env[62952]: DEBUG nova.network.neutron [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.602978] env[62952]: INFO nova.compute.claims [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1181.684780] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.883556] env[62952]: DEBUG nova.network.neutron [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updated VIF entry in instance network info cache for port de8307b9-6263-4138-befe-40f603f1bea4. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1181.883936] env[62952]: DEBUG nova.network.neutron [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.107907] env[62952]: INFO nova.compute.resource_tracker [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating resource usage from migration 63e613a9-4c2d-46b9-9284-831106c46064 [ 1182.156030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.156030] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.187736] env[62952]: DEBUG nova.compute.manager [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62952) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1182.211715] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75de74c1-94f7-440a-9871-09cd5a418f05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.219288] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6ff69b-302d-4f6f-a8cd-48174bb49fbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.252661] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c796f4cc-aa65-42ce-8084-acb0e9566107 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.260508] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269a19e1-baa5-44db-bec2-a977dccdac0d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.274209] env[62952]: DEBUG nova.compute.provider_tree [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1182.386262] env[62952]: DEBUG oslo_concurrency.lockutils [req-272dc0ed-db23-432a-807a-1f2f5dec4cf0 req-ae7be50c-d856-4782-8a3b-03ed00aa3e80 service nova] Releasing lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.662126] env[62952]: DEBUG nova.compute.utils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1182.777033] env[62952]: DEBUG nova.scheduler.client.report [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1183.164937] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.282600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.185s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.282903] env[62952]: INFO nova.compute.manager [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Migrating [ 1183.299036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.299036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.801096] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.801297] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.801392] env[62952]: DEBUG nova.network.neutron [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1183.803018] env[62952]: DEBUG nova.objects.instance [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'migration_context' on Instance uuid ff8301b2-c5cb-4e46-8f5c-98577dfd4835 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.200189] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.257825] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.258171] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.258441] env[62952]: INFO nova.compute.manager [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Attaching volume 7d69d7ff-43b2-4df1-b92c-dcc492261784 to /dev/sdb [ 1184.293689] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656b56ad-c95b-4c7c-9311-cd87d67f91ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.300917] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a8dada-3b73-484c-aa4e-9fb3ba4bde3a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.314699] env[62952]: DEBUG nova.virt.block_device [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updating existing volume attachment record: ba5d1ea8-8505-46ab-96f0-6f9d132d3e40 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1184.422767] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8414a466-71bb-4f41-96f7-5ff12bb88a7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.428676] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352c0c4b-1c56-420a-8888-a97f72c879bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.462351] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf14e1b4-eea6-4977-85d4-5cb7a3245c0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.469955] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d6f711-7ab8-44e7-a382-b3cba7393db9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.487609] env[62952]: DEBUG nova.compute.provider_tree [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.568389] env[62952]: DEBUG nova.network.neutron [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.704929] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Getting list of instances from cluster (obj){ [ 1184.704929] env[62952]: value = "domain-c8" [ 1184.704929] env[62952]: _type = "ClusterComputeResource" [ 1184.704929] env[62952]: } {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1184.705932] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43da32a-a3ee-4dae-9179-8b0cdf40155a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.724728] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Got total of 6 instances {{(pid=62952) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1184.724958] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Triggering sync for uuid aef28168-98a7-4f65-80e7-731633339abf {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1184.725196] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Triggering sync for uuid ff8301b2-c5cb-4e46-8f5c-98577dfd4835 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1184.725396] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Triggering sync for uuid 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1184.725585] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Triggering sync for uuid 26fdfac2-4ea1-4822-b38c-d8f7388436be {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1184.725771] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Triggering sync for uuid 6606e84c-36cd-4519-beff-096a8f1bc09e {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1184.725998] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Triggering sync for uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1184.726504] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "aef28168-98a7-4f65-80e7-731633339abf" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.726800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "aef28168-98a7-4f65-80e7-731633339abf" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.727136] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.727399] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.727680] env[62952]: INFO nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1184.727900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.728226] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.728447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.728789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.729114] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.729357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.729543] env[62952]: INFO nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1184.729740] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.730117] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.730433] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.731776] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11dcb60-a4e4-4871-b007-d7ba3319eec9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.736576] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdee153-afee-4597-a145-3b6d5b89be90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.739984] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce650ce6-3dd8-4706-a6e7-642d43fa497b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.991327] env[62952]: DEBUG nova.scheduler.client.report [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1185.070882] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.254614] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "aef28168-98a7-4f65-80e7-731633339abf" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.255152] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.256587] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.002976] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.703s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.584284] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da76973-6db4-48a1-9acc-8c15ed20f37e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.602963] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 0 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1187.109029] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.109285] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c832298d-32fe-45e0-b09e-f2be323fc58f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.116808] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1187.116808] env[62952]: value = "task-1367778" [ 1187.116808] env[62952]: _type = "Task" [ 1187.116808] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.123969] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.543039] env[62952]: INFO nova.compute.manager [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Swapping old allocation on dict_keys(['9b9257d4-3be7-42a5-ae8c-67cf64823681']) held by migration de73d5a1-7409-4413-b8be-1f26c2c2ff67 for instance [ 1187.564486] env[62952]: DEBUG nova.scheduler.client.report [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Overwriting current allocation {'allocations': {'9b9257d4-3be7-42a5-ae8c-67cf64823681': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 144}}, 'project_id': '0c451895447b473eb18675f6bc843081', 'user_id': '8171575c58594e7ebc37af7c4c858c10', 'consumer_generation': 1} on consumer ff8301b2-c5cb-4e46-8f5c-98577dfd4835 {{(pid=62952) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1187.627020] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367778, 'name': PowerOffVM_Task, 'duration_secs': 0.192596} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.627020] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1187.627020] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 17 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1187.643494] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.643668] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.643843] env[62952]: DEBUG nova.network.neutron [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1188.132559] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1188.132838] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1188.132970] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1188.133181] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1188.133329] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1188.133477] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1188.133679] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1188.133839] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1188.134014] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1188.134194] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1188.134372] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1188.139432] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-500c5e20-b5e4-4d78-b476-4b067201836e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.157075] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1188.157075] env[62952]: value = "task-1367779" [ 1188.157075] env[62952]: _type = "Task" [ 1188.157075] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.164975] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.367091] env[62952]: DEBUG nova.network.neutron [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [{"id": "612c5fea-4183-448f-9326-6ed1888b591d", "address": "fa:16:3e:4e:9c:c0", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap612c5fea-41", "ovs_interfaceid": "612c5fea-4183-448f-9326-6ed1888b591d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.667188] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367779, 'name': ReconfigVM_Task, 'duration_secs': 0.182716} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.667529] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 33 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1188.859532] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1188.859801] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290997', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'name': 'volume-7d69d7ff-43b2-4df1-b92c-dcc492261784', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '26fdfac2-4ea1-4822-b38c-d8f7388436be', 'attached_at': '', 'detached_at': '', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'serial': '7d69d7ff-43b2-4df1-b92c-dcc492261784'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1188.860918] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ce756c-e694-406e-9ccd-aa3b60e6099a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.878808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-ff8301b2-c5cb-4e46-8f5c-98577dfd4835" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.879752] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4733205-f702-4db2-aaee-e502ac84ebca {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.882670] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f27380-42ec-4422-bd82-b2984a429081 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.907651] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] volume-7d69d7ff-43b2-4df1-b92c-dcc492261784/volume-7d69d7ff-43b2-4df1-b92c-dcc492261784.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.908985] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db672916-9590-42f7-8277-1ee0e76bdf1e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.922117] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1973e2f8-1f08-480c-95c7-c28af11b1686 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.930464] env[62952]: DEBUG oslo_vmware.api [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1188.930464] env[62952]: value = "task-1367780" [ 1188.930464] env[62952]: _type = "Task" [ 1188.930464] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.943632] env[62952]: DEBUG oslo_vmware.api [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367780, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.173878] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1189.174113] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1189.174271] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1189.174459] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1189.174612] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1189.174767] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1189.174975] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1189.175166] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1189.175339] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1189.175505] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1189.175684] env[62952]: DEBUG nova.virt.hardware [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1189.181230] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1189.181525] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7b50a74-1f14-4ddd-bd4f-c1f49f2f1ef1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.200317] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1189.200317] env[62952]: value = "task-1367781" [ 1189.200317] env[62952]: _type = "Task" [ 1189.200317] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.207810] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367781, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.439724] env[62952]: DEBUG oslo_vmware.api [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367780, 'name': ReconfigVM_Task, 'duration_secs': 0.326128} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.439967] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Reconfigured VM instance instance-0000006a to attach disk [datastore2] volume-7d69d7ff-43b2-4df1-b92c-dcc492261784/volume-7d69d7ff-43b2-4df1-b92c-dcc492261784.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.444983] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9384dd4-624a-4841-a47a-3e289caa7493 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.459577] env[62952]: DEBUG oslo_vmware.api [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1189.459577] env[62952]: value = "task-1367782" [ 1189.459577] env[62952]: _type = "Task" [ 1189.459577] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.466996] env[62952]: DEBUG oslo_vmware.api [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367782, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.712140] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367781, 'name': ReconfigVM_Task, 'duration_secs': 0.155862} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.712480] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1189.713198] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e8aeeb-fcf7-4a78-843a-a10e7fd441b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.734358] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-188d94f2-707a-48de-b891-ba671ade4185/volume-188d94f2-707a-48de-b891-ba671ade4185.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1189.734571] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53ec1f6e-ef3d-4c0b-a450-eb7f914f0838 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.751511] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1189.751511] env[62952]: value = "task-1367783" [ 1189.751511] env[62952]: _type = "Task" [ 1189.751511] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.758706] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367783, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.969188] env[62952]: DEBUG oslo_vmware.api [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367782, 'name': ReconfigVM_Task, 'duration_secs': 0.130618} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.969486] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290997', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'name': 'volume-7d69d7ff-43b2-4df1-b92c-dcc492261784', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '26fdfac2-4ea1-4822-b38c-d8f7388436be', 'attached_at': '', 'detached_at': '', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'serial': '7d69d7ff-43b2-4df1-b92c-dcc492261784'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1190.008042] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1190.008354] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca6c625e-c36b-48ef-b796-52089e34482b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.017165] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1190.017165] env[62952]: value = "task-1367784" [ 1190.017165] env[62952]: _type = "Task" [ 1190.017165] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.024395] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.261595] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367783, 'name': ReconfigVM_Task, 'duration_secs': 0.233204} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.261840] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-188d94f2-707a-48de-b891-ba671ade4185/volume-188d94f2-707a-48de-b891-ba671ade4185.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1190.262080] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 50 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1190.527381] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.769300] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249cab6b-6cee-4d94-a0b0-e0b18bd41cbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.787993] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b688c02-87e5-4714-817f-58509e071ef8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.805207] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 67 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1191.006926] env[62952]: DEBUG nova.objects.instance [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'flavor' on Instance uuid 26fdfac2-4ea1-4822-b38c-d8f7388436be {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.029035] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.511490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-77f22619-92c0-42b2-88e0-402edc980cb2 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.253s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.512371] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.784s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.513363] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3064f15c-90ce-447d-a629-180d6b45d250 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.530259] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367784, 'name': PowerOffVM_Task, 'duration_secs': 1.115666} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.530521] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1191.531141] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.531357] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.531513] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.531696] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.531844] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.531993] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.532267] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.532463] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.532635] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.532799] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.532975] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.538316] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5637fd34-90f4-4906-9301-cf8ed1852092 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.552859] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1191.552859] env[62952]: value = "task-1367785" [ 1191.552859] env[62952]: _type = "Task" [ 1191.552859] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.559974] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.735746] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.025748] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.513s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.026104] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.290s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.062765] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367785, 'name': ReconfigVM_Task, 'duration_secs': 0.137378} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.063571] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086f526a-21de-4169-a285-5b0ae181d9a0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.084028] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1192.084274] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1192.084436] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1192.084623] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1192.084775] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1192.084934] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1192.085202] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1192.085330] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1192.085491] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1192.085669] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1192.085876] env[62952]: DEBUG nova.virt.hardware [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1192.086637] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0dad84e-0d4a-4abf-a0ad-debf9297357e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.091918] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1192.091918] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]525df56b-88f5-a5be-550f-b312d33f1e18" [ 1192.091918] env[62952]: _type = "Task" [ 1192.091918] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.099827] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525df56b-88f5-a5be-550f-b312d33f1e18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.528911] env[62952]: INFO nova.compute.manager [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Detaching volume 7d69d7ff-43b2-4df1-b92c-dcc492261784 [ 1192.551501] env[62952]: DEBUG nova.network.neutron [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Port de8307b9-6263-4138-befe-40f603f1bea4 binding to destination host cpu-1 is already ACTIVE {{(pid=62952) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1192.558857] env[62952]: INFO nova.virt.block_device [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Attempting to driver detach volume 7d69d7ff-43b2-4df1-b92c-dcc492261784 from mountpoint /dev/sdb [ 1192.559164] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1192.559373] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290997', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'name': 'volume-7d69d7ff-43b2-4df1-b92c-dcc492261784', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '26fdfac2-4ea1-4822-b38c-d8f7388436be', 'attached_at': '', 'detached_at': '', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'serial': '7d69d7ff-43b2-4df1-b92c-dcc492261784'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1192.560226] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a68f154-0d69-45ff-9c70-9f9b0ccbf985 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.581966] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830a3cce-c71c-4786-a7a9-1b3d371f2b1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.588619] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f30d660-4333-4784-bec1-ac4798b74903 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.613620] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1b4c6f-b49a-4d2a-b7cc-f3d94080c4c3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.619122] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]525df56b-88f5-a5be-550f-b312d33f1e18, 'name': SearchDatastore_Task, 'duration_secs': 0.010628} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.624581] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1192.634433] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e53250bf-4c87-453e-ba08-ae9e4b23fb0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.646822] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] The volume has not been displaced from its original location: [datastore2] volume-7d69d7ff-43b2-4df1-b92c-dcc492261784/volume-7d69d7ff-43b2-4df1-b92c-dcc492261784.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1192.651923] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1192.652203] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38230364-c589-4d47-b494-e25677de0646 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.668855] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1192.668855] env[62952]: value = "task-1367786" [ 1192.668855] env[62952]: _type = "Task" [ 1192.668855] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.670021] env[62952]: DEBUG oslo_vmware.api [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1192.670021] env[62952]: value = "task-1367787" [ 1192.670021] env[62952]: _type = "Task" [ 1192.670021] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.680171] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367786, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.683031] env[62952]: DEBUG oslo_vmware.api [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367787, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.184759] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367786, 'name': ReconfigVM_Task, 'duration_secs': 0.205744} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.185169] env[62952]: DEBUG oslo_vmware.api [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367787, 'name': ReconfigVM_Task, 'duration_secs': 0.248735} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.185222] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1193.185515] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1193.190553] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfdcd5c-8d8b-4845-b9d4-00fdc7c9e4d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.192976] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c98aea20-f9e4-4cf1-8adb-5a79dc3fed17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.224132] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1193.225418] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60a94ecf-85b3-485b-ba4b-7d541af0c47e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.237905] env[62952]: DEBUG oslo_vmware.api [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1193.237905] env[62952]: value = "task-1367788" [ 1193.237905] env[62952]: _type = "Task" [ 1193.237905] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.244529] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1193.244529] env[62952]: value = "task-1367789" [ 1193.244529] env[62952]: _type = "Task" [ 1193.244529] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.250427] env[62952]: DEBUG oslo_vmware.api [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.255206] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367789, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.572675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.572918] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.573084] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.747942] env[62952]: DEBUG oslo_vmware.api [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367788, 'name': ReconfigVM_Task, 'duration_secs': 0.14811} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.750851] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290997', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'name': 'volume-7d69d7ff-43b2-4df1-b92c-dcc492261784', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '26fdfac2-4ea1-4822-b38c-d8f7388436be', 'attached_at': '', 'detached_at': '', 'volume_id': '7d69d7ff-43b2-4df1-b92c-dcc492261784', 'serial': '7d69d7ff-43b2-4df1-b92c-dcc492261784'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1193.757341] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367789, 'name': ReconfigVM_Task, 'duration_secs': 0.28964} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.757527] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to attach disk [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835/ff8301b2-c5cb-4e46-8f5c-98577dfd4835.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1193.758328] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c55291-db19-4865-85ff-b02ade88d9d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.778175] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2b7d5e-1995-4d98-8fff-78860ee81e28 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.798276] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d6fc15-8afa-4d2c-a505-be0958076e81 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.819480] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd100f9c-713a-45c0-9aac-e5bd9d4307bd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.826805] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1193.827104] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75933a1c-c897-47e5-ab59-bbe807914ce5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.833644] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1193.833644] env[62952]: value = "task-1367790" [ 1193.833644] env[62952]: _type = "Task" [ 1193.833644] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.841070] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.290189] env[62952]: DEBUG nova.objects.instance [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'flavor' on Instance uuid 26fdfac2-4ea1-4822-b38c-d8f7388436be {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.345473] env[62952]: DEBUG oslo_vmware.api [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367790, 'name': PowerOnVM_Task, 'duration_secs': 0.392812} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.346096] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1194.670989] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.671428] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.671428] env[62952]: DEBUG nova.network.neutron [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1195.297505] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea9a7ea1-536b-410a-8416-18e28247c6fa tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.271s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.372494] env[62952]: DEBUG nova.network.neutron [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.387031] env[62952]: INFO nova.compute.manager [None req-6046cee7-0e85-486f-a3dd-ccae11a05b21 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance to original state: 'active' [ 1195.876070] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.967496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.967496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.967671] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "26fdfac2-4ea1-4822-b38c-d8f7388436be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.967808] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.967993] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.969993] env[62952]: INFO nova.compute.manager [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Terminating instance [ 1195.971774] env[62952]: DEBUG nova.compute.manager [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1195.971974] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1195.972814] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c41eb42-ca79-482a-a44a-6058249aa198 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.980633] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1195.980857] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35f1ad8e-ea88-40ee-a75b-c224420ea3d9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.987765] env[62952]: DEBUG oslo_vmware.api [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1195.987765] env[62952]: value = "task-1367791" [ 1195.987765] env[62952]: _type = "Task" [ 1195.987765] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.995256] env[62952]: DEBUG oslo_vmware.api [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.319295] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.319577] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.319770] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.320011] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.320207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.322351] env[62952]: INFO nova.compute.manager [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Terminating instance [ 1196.324275] env[62952]: DEBUG nova.compute.manager [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1196.324489] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.324733] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93ce573d-f271-446f-92e7-c3d8cd719c8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.331016] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1196.331016] env[62952]: value = "task-1367792" [ 1196.331016] env[62952]: _type = "Task" [ 1196.331016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.340875] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.384635] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79315d15-882b-4a2e-8014-c235b51fa365 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.391307] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a000fcc1-4882-41b0-8838-62cb47738c52 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.497319] env[62952]: DEBUG oslo_vmware.api [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367791, 'name': PowerOffVM_Task, 'duration_secs': 0.208199} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.497680] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.497901] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.498224] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e062316d-bd84-41ad-a71a-3ba16f6bdca7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.559720] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.559965] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.560139] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleting the datastore file [datastore1] 26fdfac2-4ea1-4822-b38c-d8f7388436be {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.560408] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-712e7eae-d3cd-47c9-a515-bc8d24089783 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.566497] env[62952]: DEBUG oslo_vmware.api [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1196.566497] env[62952]: value = "task-1367794" [ 1196.566497] env[62952]: _type = "Task" [ 1196.566497] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.573847] env[62952]: DEBUG oslo_vmware.api [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367794, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.840778] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367792, 'name': PowerOffVM_Task, 'duration_secs': 0.197737} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.840995] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.841216] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1196.841412] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290991', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'name': 'volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835', 'attached_at': '2024-10-26T01:43:55.000000', 'detached_at': '', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'serial': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1196.842166] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443cedf0-e75d-4987-bae5-699ee204b988 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.862352] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d9b111-ecdf-4af1-b449-5bb97abb5150 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.868358] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523dc748-0e47-481a-8d97-e23e29d80e6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.887381] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163ee5a1-6294-4491-8439-a181a8aaca51 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.901217] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] The volume has not been displaced from its original location: [datastore1] volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6/volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1196.906342] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1196.906602] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34f44af3-5be3-4aea-a93e-6bffa0bf1144 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.923115] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1196.923115] env[62952]: value = "task-1367795" [ 1196.923115] env[62952]: _type = "Task" [ 1196.923115] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.932754] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367795, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.076266] env[62952]: DEBUG oslo_vmware.api [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367794, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147995} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.076528] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.076718] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1197.076901] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1197.077096] env[62952]: INFO nova.compute.manager [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1197.077345] env[62952]: DEBUG oslo.service.loopingcall [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1197.077537] env[62952]: DEBUG nova.compute.manager [-] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1197.077654] env[62952]: DEBUG nova.network.neutron [-] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1197.433284] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367795, 'name': ReconfigVM_Task, 'duration_secs': 0.214447} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.433770] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1197.438114] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc012d6a-da49-477d-8169-f3a77bb2d766 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.453859] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1197.453859] env[62952]: value = "task-1367796" [ 1197.453859] env[62952]: _type = "Task" [ 1197.453859] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.464718] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367796, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.494445] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3bac5b-8cf8-4fd9-b582-641b05ec4558 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.515278] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fcc943-8e85-420d-bf3c-3b0d33fc0a60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.522403] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 83 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1197.536900] env[62952]: DEBUG nova.compute.manager [req-7ffc2812-3989-421c-a764-423f6b06badd req-e522d6bc-8895-4ff0-905c-5d150e4e2dab service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Received event network-vif-deleted-c00369d7-037b-4a8e-9f3e-cd130492b005 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1197.537177] env[62952]: INFO nova.compute.manager [req-7ffc2812-3989-421c-a764-423f6b06badd req-e522d6bc-8895-4ff0-905c-5d150e4e2dab service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Neutron deleted interface c00369d7-037b-4a8e-9f3e-cd130492b005; detaching it from the instance and deleting it from the info cache [ 1197.537311] env[62952]: DEBUG nova.network.neutron [req-7ffc2812-3989-421c-a764-423f6b06badd req-e522d6bc-8895-4ff0-905c-5d150e4e2dab service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.964919] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367796, 'name': ReconfigVM_Task, 'duration_secs': 0.131933} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.964919] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290991', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'name': 'volume-ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ff8301b2-c5cb-4e46-8f5c-98577dfd4835', 'attached_at': '2024-10-26T01:43:55.000000', 'detached_at': '', 'volume_id': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6', 'serial': 'ecd5ca41-746f-4f21-b85b-ae74dec11fc6'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1197.964919] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1197.965636] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b15575-de7b-43d0-9b2b-59fed1dd5fcb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.971968] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1197.972270] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b84ea6f-d9b2-492f-9cbd-2a5c5781ff27 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.019182] env[62952]: DEBUG nova.network.neutron [-] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.028549] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1198.029397] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6cfb2c2-d62f-45f0-aca5-1115effed052 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.037273] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1198.037273] env[62952]: value = "task-1367798" [ 1198.037273] env[62952]: _type = "Task" [ 1198.037273] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.041173] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-582101ab-9992-4e6b-9c3a-5ff4e793ebc3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.047944] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.052346] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3280f3-a1f3-4619-b0ad-3ac1754eb261 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.066388] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1198.066614] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1198.066801] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleting the datastore file [datastore1] ff8301b2-c5cb-4e46-8f5c-98577dfd4835 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.067449] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b856719-1491-4250-94fc-20b794541550 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.074016] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1198.074016] env[62952]: value = "task-1367799" [ 1198.074016] env[62952]: _type = "Task" [ 1198.074016] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.084135] env[62952]: DEBUG nova.compute.manager [req-7ffc2812-3989-421c-a764-423f6b06badd req-e522d6bc-8895-4ff0-905c-5d150e4e2dab service nova] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Detach interface failed, port_id=c00369d7-037b-4a8e-9f3e-cd130492b005, reason: Instance 26fdfac2-4ea1-4822-b38c-d8f7388436be could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1198.090378] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367799, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.521765] env[62952]: INFO nova.compute.manager [-] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Took 1.44 seconds to deallocate network for instance. [ 1198.547324] env[62952]: DEBUG oslo_vmware.api [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367798, 'name': PowerOnVM_Task, 'duration_secs': 0.378778} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.547727] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1198.547990] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9c1a4a-27e8-42ed-b05a-680db388293b tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance '6606e84c-36cd-4519-beff-096a8f1bc09e' progress to 100 {{(pid=62952) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1198.583716] env[62952]: DEBUG oslo_vmware.api [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367799, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150695} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.583962] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.584166] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1198.584347] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1198.584526] env[62952]: INFO nova.compute.manager [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1198.584769] env[62952]: DEBUG oslo.service.loopingcall [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.585028] env[62952]: DEBUG nova.compute.manager [-] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1198.585145] env[62952]: DEBUG nova.network.neutron [-] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1199.029801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.029801] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.029801] env[62952]: DEBUG nova.objects.instance [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'resources' on Instance uuid 26fdfac2-4ea1-4822-b38c-d8f7388436be {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.560349] env[62952]: DEBUG nova.compute.manager [req-9d44c64b-4ef1-4ba4-bc03-272b1fff5cee req-7ac37ab9-9ea2-4f7d-972f-122683309189 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Received event network-vif-deleted-612c5fea-4183-448f-9326-6ed1888b591d {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.560622] env[62952]: INFO nova.compute.manager [req-9d44c64b-4ef1-4ba4-bc03-272b1fff5cee req-7ac37ab9-9ea2-4f7d-972f-122683309189 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Neutron deleted interface 612c5fea-4183-448f-9326-6ed1888b591d; detaching it from the instance and deleting it from the info cache [ 1199.560914] env[62952]: DEBUG nova.network.neutron [req-9d44c64b-4ef1-4ba4-bc03-272b1fff5cee req-7ac37ab9-9ea2-4f7d-972f-122683309189 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.635696] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d215ea71-461a-48de-9e14-2719d32124a4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.643996] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3b889b-2fbf-4a7f-ab88-3cc89643ad7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.674172] env[62952]: DEBUG nova.network.neutron [-] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.676792] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6173bdd-9f19-4009-b2f1-e56180cf3e43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.683805] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e608d53-b227-47d6-819e-bad13a0ce57b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.697526] env[62952]: DEBUG nova.compute.provider_tree [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.063366] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90308f35-93d1-499d-96ab-46ebe0314dc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.072488] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b71bfb-8c9a-447b-8a3e-553251d4732c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.102449] env[62952]: DEBUG nova.compute.manager [req-9d44c64b-4ef1-4ba4-bc03-272b1fff5cee req-7ac37ab9-9ea2-4f7d-972f-122683309189 service nova] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Detach interface failed, port_id=612c5fea-4183-448f-9326-6ed1888b591d, reason: Instance ff8301b2-c5cb-4e46-8f5c-98577dfd4835 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1200.178805] env[62952]: INFO nova.compute.manager [-] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Took 1.59 seconds to deallocate network for instance. [ 1200.200389] env[62952]: DEBUG nova.scheduler.client.report [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1200.704973] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.724946] env[62952]: INFO nova.compute.manager [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Took 0.55 seconds to detach 1 volumes for instance. [ 1200.730863] env[62952]: INFO nova.scheduler.client.report [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted allocations for instance 26fdfac2-4ea1-4822-b38c-d8f7388436be [ 1201.216549] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.216863] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.217081] env[62952]: DEBUG nova.compute.manager [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Going to confirm migration 6 {{(pid=62952) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1201.231317] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.231572] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.231769] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.236669] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ee03654c-935a-4343-9894-1674ea02cc76 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "26fdfac2-4ea1-4822-b38c-d8f7388436be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.269s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.255063] env[62952]: INFO nova.scheduler.client.report [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted allocations for instance ff8301b2-c5cb-4e46-8f5c-98577dfd4835 [ 1201.753796] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.754075] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquired lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.754180] env[62952]: DEBUG nova.network.neutron [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1201.754367] env[62952]: DEBUG nova.objects.instance [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'info_cache' on Instance uuid 6606e84c-36cd-4519-beff-096a8f1bc09e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.762851] env[62952]: DEBUG oslo_concurrency.lockutils [None req-1f317f12-c805-461e-93d5-6aa484a970d5 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "ff8301b2-c5cb-4e46-8f5c-98577dfd4835" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.443s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.782214] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.782483] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.983436] env[62952]: DEBUG nova.network.neutron [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [{"id": "de8307b9-6263-4138-befe-40f603f1bea4", "address": "fa:16:3e:e8:94:07", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde8307b9-62", "ovs_interfaceid": "de8307b9-6263-4138-befe-40f603f1bea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.284775] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1203.366411] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.366652] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.486311] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Releasing lock "refresh_cache-6606e84c-36cd-4519-beff-096a8f1bc09e" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.486518] env[62952]: DEBUG nova.objects.instance [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'migration_context' on Instance uuid 6606e84c-36cd-4519-beff-096a8f1bc09e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.804206] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.804512] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.806137] env[62952]: INFO nova.compute.claims [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1203.868963] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1203.989483] env[62952]: DEBUG nova.objects.base [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Object Instance<6606e84c-36cd-4519-beff-096a8f1bc09e> lazy-loaded attributes: info_cache,migration_context {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1203.990474] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95511dc-4995-4b37-8797-3027f244e5db {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.010849] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e1cadf9-7bf4-4652-8bf4-9570cd8834e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.015872] env[62952]: DEBUG oslo_vmware.api [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1204.015872] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52cc622f-b70e-84d9-af73-4e5a0fcbe2e5" [ 1204.015872] env[62952]: _type = "Task" [ 1204.015872] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.023515] env[62952]: DEBUG oslo_vmware.api [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52cc622f-b70e-84d9-af73-4e5a0fcbe2e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.390640] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.524964] env[62952]: DEBUG oslo_vmware.api [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52cc622f-b70e-84d9-af73-4e5a0fcbe2e5, 'name': SearchDatastore_Task, 'duration_secs': 0.009614} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.525251] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.895708] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf66042d-923e-41b9-9193-13f79528d095 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.903517] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e1e561-f1e4-49ef-9d94-b337d6a30fd5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.932354] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f61111-7c59-455f-b545-a621d8e74609 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.939648] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbe60a6-ecb8-41f1-8820-07a727a364bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.953689] env[62952]: DEBUG nova.compute.provider_tree [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.457122] env[62952]: DEBUG nova.scheduler.client.report [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1205.714505] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.714706] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.714885] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.962109] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.962669] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1205.965424] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.575s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.966811] env[62952]: INFO nova.compute.claims [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1206.217600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.471410] env[62952]: DEBUG nova.compute.utils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1206.475051] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1206.475051] env[62952]: DEBUG nova.network.neutron [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1206.511711] env[62952]: DEBUG nova.policy [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e67193d02201461bb7e4339b5025ca48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfc390d64c0463190f071f3f62936dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1206.755579] env[62952]: DEBUG nova.network.neutron [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Successfully created port: 199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1206.975572] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1207.077993] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e0d334-bccb-4128-b7df-a8b718f32f61 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.085453] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cc6071-d9c7-43ba-8fb6-fc14308c5634 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.126703] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dfbfc6-54bf-4d54-933a-54444e384223 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.136084] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2661eaa2-d121-4cc8-89b8-fe5f45ae8189 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.154302] env[62952]: DEBUG nova.compute.provider_tree [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.657935] env[62952]: DEBUG nova.scheduler.client.report [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1207.987048] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1208.012267] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1208.012529] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1208.012692] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1208.012877] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1208.013346] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1208.013346] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1208.013468] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1208.013541] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1208.013706] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1208.013873] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1208.014064] env[62952]: DEBUG nova.virt.hardware [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1208.014932] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799b2d85-5727-4b12-8d9a-7f1047054bd0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.022627] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72be215c-ccc5-4918-adae-b1e83ce30d08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.140704] env[62952]: DEBUG nova.compute.manager [req-821986c2-3a18-494f-9c6d-a222b3077e05 req-78e123a4-3fc5-4550-9d77-28d143b9e1da service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Received event network-vif-plugged-199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1208.140967] env[62952]: DEBUG oslo_concurrency.lockutils [req-821986c2-3a18-494f-9c6d-a222b3077e05 req-78e123a4-3fc5-4550-9d77-28d143b9e1da service nova] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.141210] env[62952]: DEBUG oslo_concurrency.lockutils [req-821986c2-3a18-494f-9c6d-a222b3077e05 req-78e123a4-3fc5-4550-9d77-28d143b9e1da service nova] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.141404] env[62952]: DEBUG oslo_concurrency.lockutils [req-821986c2-3a18-494f-9c6d-a222b3077e05 req-78e123a4-3fc5-4550-9d77-28d143b9e1da service nova] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.141559] env[62952]: DEBUG nova.compute.manager [req-821986c2-3a18-494f-9c6d-a222b3077e05 req-78e123a4-3fc5-4550-9d77-28d143b9e1da service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] No waiting events found dispatching network-vif-plugged-199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1208.141723] env[62952]: WARNING nova.compute.manager [req-821986c2-3a18-494f-9c6d-a222b3077e05 req-78e123a4-3fc5-4550-9d77-28d143b9e1da service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Received unexpected event network-vif-plugged-199c0f83-9683-4568-806b-f01b051f2f0c for instance with vm_state building and task_state spawning. [ 1208.162966] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.198s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.163471] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1208.166065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.641s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.227623] env[62952]: DEBUG nova.network.neutron [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Successfully updated port: 199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1208.669421] env[62952]: DEBUG nova.compute.utils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1208.673863] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1208.674149] env[62952]: DEBUG nova.network.neutron [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1208.714875] env[62952]: DEBUG nova.policy [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8171575c58594e7ebc37af7c4c858c10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c451895447b473eb18675f6bc843081', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1208.732962] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1208.733126] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.734918] env[62952]: DEBUG nova.network.neutron [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1208.770224] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5cc2b7-6e7b-4029-a987-f5d60b3fd544 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.779604] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc318ee-afb4-4e0a-9ea0-adbacbe42dfc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.809551] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952ea1af-0602-4c01-96d5-c4ef2e268451 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.816917] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0777079-8730-430a-8e87-98572a6da5ec {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.830343] env[62952]: DEBUG nova.compute.provider_tree [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.972280] env[62952]: DEBUG nova.network.neutron [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Successfully created port: 208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1209.174506] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1209.262499] env[62952]: DEBUG nova.network.neutron [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1209.332887] env[62952]: DEBUG nova.scheduler.client.report [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1209.389120] env[62952]: DEBUG nova.network.neutron [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updating instance_info_cache with network_info: [{"id": "199c0f83-9683-4568-806b-f01b051f2f0c", "address": "fa:16:3e:75:e1:0d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap199c0f83-96", "ovs_interfaceid": "199c0f83-9683-4568-806b-f01b051f2f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.891340] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1209.891637] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Instance network_info: |[{"id": "199c0f83-9683-4568-806b-f01b051f2f0c", "address": "fa:16:3e:75:e1:0d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap199c0f83-96", "ovs_interfaceid": "199c0f83-9683-4568-806b-f01b051f2f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1209.892073] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:e1:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '199c0f83-9683-4568-806b-f01b051f2f0c', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1209.899716] env[62952]: DEBUG oslo.service.loopingcall [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1209.899918] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1209.900155] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e22cdc0-b93c-427f-a811-c43ae8bf6dcc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.920679] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1209.920679] env[62952]: value = "task-1367800" [ 1209.920679] env[62952]: _type = "Task" [ 1209.920679] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.928074] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367800, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.170868] env[62952]: DEBUG nova.compute.manager [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Received event network-changed-199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1210.171112] env[62952]: DEBUG nova.compute.manager [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Refreshing instance network info cache due to event network-changed-199c0f83-9683-4568-806b-f01b051f2f0c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1210.171342] env[62952]: DEBUG oslo_concurrency.lockutils [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] Acquiring lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.171491] env[62952]: DEBUG oslo_concurrency.lockutils [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] Acquired lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.171663] env[62952]: DEBUG nova.network.neutron [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Refreshing network info cache for port 199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1210.183788] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1210.210373] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1210.210626] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1210.210786] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1210.210971] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1210.211137] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1210.211333] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1210.211499] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1210.211663] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1210.211834] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1210.211998] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1210.212216] env[62952]: DEBUG nova.virt.hardware [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1210.213083] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe0820f-a96f-44b0-948b-344cf2bffb2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.221202] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1805cdb-7274-4047-b84b-233a998ee939 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.343887] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.178s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.346712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.129s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.346898] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.347065] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1210.348167] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b857ed5-e34f-439c-af3d-80eb83c97403 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.356272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41801cff-ba4a-445c-9af2-e69755cbf2d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.369614] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3f5567-83a5-4431-b90a-60abcf174d2c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.375663] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b70c9b-61e4-4603-b3c1-3dbaddfc8a0b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.405576] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180694MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1210.405687] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.405858] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.417910] env[62952]: DEBUG nova.network.neutron [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Successfully updated port: 208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1210.429849] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367800, 'name': CreateVM_Task, 'duration_secs': 0.344766} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.430010] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1210.430683] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.430847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.431211] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1210.431689] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5da3ece8-c73a-4572-915b-29195b64122a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.437331] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1210.437331] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52719c4a-46d8-74aa-b783-f709284c23d5" [ 1210.437331] env[62952]: _type = "Task" [ 1210.437331] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.445565] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52719c4a-46d8-74aa-b783-f709284c23d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.914253] env[62952]: INFO nova.scheduler.client.report [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted allocation for migration 63e613a9-4c2d-46b9-9284-831106c46064 [ 1210.920036] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.920188] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.920333] env[62952]: DEBUG nova.network.neutron [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1210.947790] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52719c4a-46d8-74aa-b783-f709284c23d5, 'name': SearchDatastore_Task, 'duration_secs': 0.009884} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.948143] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.948424] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1210.948663] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.948816] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.948999] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1210.949276] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4816e909-2d65-4e94-ad9f-636f92baa6ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.951886] env[62952]: DEBUG nova.network.neutron [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updated VIF entry in instance network info cache for port 199c0f83-9683-4568-806b-f01b051f2f0c. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1210.952242] env[62952]: DEBUG nova.network.neutron [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updating instance_info_cache with network_info: [{"id": "199c0f83-9683-4568-806b-f01b051f2f0c", "address": "fa:16:3e:75:e1:0d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap199c0f83-96", "ovs_interfaceid": "199c0f83-9683-4568-806b-f01b051f2f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.960487] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1210.960691] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1210.961701] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33ee822c-38e9-4284-bad2-89bf1bdaabdc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.967390] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1210.967390] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]529b09ee-a7de-d3a4-5c62-78890b8e75f7" [ 1210.967390] env[62952]: _type = "Task" [ 1210.967390] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.975308] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529b09ee-a7de-d3a4-5c62-78890b8e75f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.983079] env[62952]: INFO nova.compute.manager [None req-206f3508-7c55-4113-a24e-b3c551fea67a tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Get console output [ 1210.983414] env[62952]: WARNING nova.virt.vmwareapi.driver [None req-206f3508-7c55-4113-a24e-b3c551fea67a tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] The console log is missing. Check your VSPC configuration [ 1211.419902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c24daf77-629a-4856-b14e-9eadf1224049 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.203s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.430957] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance aef28168-98a7-4f65-80e7-731633339abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.430957] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.431181] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 27e78e78-3a7f-4ad4-b939-438cb2b90b55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.431316] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 6606e84c-36cd-4519-beff-096a8f1bc09e actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.431592] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance dd7d45f3-31bb-43c3-9317-2142d29b6ef5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.431592] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance c1c3d578-37e0-4823-9acf-612a09f512f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.431732] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1211.431924] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1211.458729] env[62952]: DEBUG oslo_concurrency.lockutils [req-c09c15fa-155a-4471-a25a-a093dc62f979 req-f2ec2323-6cfd-49d8-96e3-8bd61dfa1e70 service nova] Releasing lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.462859] env[62952]: DEBUG nova.network.neutron [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1211.478827] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]529b09ee-a7de-d3a4-5c62-78890b8e75f7, 'name': SearchDatastore_Task, 'duration_secs': 0.008796} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.483531] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e76f072e-2f81-4183-9388-f6da5e8a53d6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.488885] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1211.488885] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52008aae-e55b-d932-cb9c-b0996a756bef" [ 1211.488885] env[62952]: _type = "Task" [ 1211.488885] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.497172] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52008aae-e55b-d932-cb9c-b0996a756bef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.522593] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cea2fd-6bed-4642-ab15-c230ba3afe9a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.529842] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06131f9-66bd-4f44-aea5-634b5b383297 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.566307] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ea8c82-7b91-4a79-bb41-fe336d4b52d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.575884] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb46df7b-1652-4850-8302-07d278d5bae2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.589108] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.602333] env[62952]: DEBUG nova.network.neutron [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.002464] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52008aae-e55b-d932-cb9c-b0996a756bef, 'name': SearchDatastore_Task, 'duration_secs': 0.009216} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.002642] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.003114] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] dd7d45f3-31bb-43c3-9317-2142d29b6ef5/dd7d45f3-31bb-43c3-9317-2142d29b6ef5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1212.003582] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ea68af7-5136-4bf2-adca-25b60f10f7ce {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.013175] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1212.013175] env[62952]: value = "task-1367801" [ 1212.013175] env[62952]: _type = "Task" [ 1212.013175] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.021467] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.092804] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1212.104592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.104989] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance network_info: |[{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1212.105597] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:fd:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '208e022a-eaea-4c27-ac52-5773ba975aa3', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1212.113742] env[62952]: DEBUG oslo.service.loopingcall [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1212.114596] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1212.114886] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3bd7914-529e-4d18-bad6-f9c39d6ef1e8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.135739] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1212.135739] env[62952]: value = "task-1367802" [ 1212.135739] env[62952]: _type = "Task" [ 1212.135739] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.143405] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367802, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.199155] env[62952]: DEBUG nova.compute.manager [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-vif-plugged-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1212.199155] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.199155] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.199155] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.199312] env[62952]: DEBUG nova.compute.manager [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] No waiting events found dispatching network-vif-plugged-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1212.199528] env[62952]: WARNING nova.compute.manager [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received unexpected event network-vif-plugged-208e022a-eaea-4c27-ac52-5773ba975aa3 for instance with vm_state building and task_state spawning. [ 1212.199708] env[62952]: DEBUG nova.compute.manager [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1212.199868] env[62952]: DEBUG nova.compute.manager [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing instance network info cache due to event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1212.200065] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.200211] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.200370] env[62952]: DEBUG nova.network.neutron [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1212.522719] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467757} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.523061] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] dd7d45f3-31bb-43c3-9317-2142d29b6ef5/dd7d45f3-31bb-43c3-9317-2142d29b6ef5.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1212.523282] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1212.523542] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-008b6fda-859f-4eb5-b917-f8d5ac5ad782 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.530785] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1212.530785] env[62952]: value = "task-1367803" [ 1212.530785] env[62952]: _type = "Task" [ 1212.530785] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.539894] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.597912] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1212.598161] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.192s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.645744] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367802, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.904864] env[62952]: DEBUG nova.network.neutron [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updated VIF entry in instance network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1212.905319] env[62952]: DEBUG nova.network.neutron [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.040315] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067192} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.040602] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1213.041394] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109c5e46-f645-42a2-82df-b05e9420c17b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.062877] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] dd7d45f3-31bb-43c3-9317-2142d29b6ef5/dd7d45f3-31bb-43c3-9317-2142d29b6ef5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1213.063159] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6079ebbd-2446-474b-b4f1-a9f1ba69f210 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.082189] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1213.082189] env[62952]: value = "task-1367804" [ 1213.082189] env[62952]: _type = "Task" [ 1213.082189] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.089730] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367804, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.145448] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367802, 'name': CreateVM_Task, 'duration_secs': 0.538797} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.145615] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1213.146305] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.146473] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.146800] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1213.147071] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb8c336d-3240-4a6a-9b17-ce33f52b7848 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.151681] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1213.151681] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5273670f-be63-08f5-fdd3-7b2bf7067c16" [ 1213.151681] env[62952]: _type = "Task" [ 1213.151681] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.159191] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5273670f-be63-08f5-fdd3-7b2bf7067c16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.408578] env[62952]: DEBUG oslo_concurrency.lockutils [req-9f24afeb-0912-44ac-9868-1e87b627bbf8 req-53c6f4bd-e639-4955-81d2-507b4fb8e88f service nova] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.592353] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367804, 'name': ReconfigVM_Task, 'duration_secs': 0.316565} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.592647] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Reconfigured VM instance instance-0000006d to attach disk [datastore2] dd7d45f3-31bb-43c3-9317-2142d29b6ef5/dd7d45f3-31bb-43c3-9317-2142d29b6ef5.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1213.593327] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38e56484-2445-43aa-a775-6f4dee5ccb48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.599975] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1213.599975] env[62952]: value = "task-1367805" [ 1213.599975] env[62952]: _type = "Task" [ 1213.599975] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.608064] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367805, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.661527] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5273670f-be63-08f5-fdd3-7b2bf7067c16, 'name': SearchDatastore_Task, 'duration_secs': 0.008321} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.661878] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.662141] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1213.662386] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.662536] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.662717] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1213.662972] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db31621a-15b8-4fa8-8bab-f7b53b16654d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.670871] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1213.671109] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1213.671873] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-598b8721-5dac-452a-9ad4-6846807060a1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.677175] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1213.677175] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279632b-a907-92d2-bf16-da14686047fc" [ 1213.677175] env[62952]: _type = "Task" [ 1213.677175] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.684208] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5279632b-a907-92d2-bf16-da14686047fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.109138] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367805, 'name': Rename_Task, 'duration_secs': 0.139132} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.109460] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1214.109733] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eacbe90f-86f2-468b-a04c-4de83ced2f68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.116551] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1214.116551] env[62952]: value = "task-1367806" [ 1214.116551] env[62952]: _type = "Task" [ 1214.116551] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.124903] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367806, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.187913] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5279632b-a907-92d2-bf16-da14686047fc, 'name': SearchDatastore_Task, 'duration_secs': 0.007783} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.188830] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb8523a4-0c67-4b95-92fb-76f32523c7b0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.193893] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1214.193893] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]520e8295-4c47-1241-79c6-018b20c3e4f8" [ 1214.193893] env[62952]: _type = "Task" [ 1214.193893] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.200871] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520e8295-4c47-1241-79c6-018b20c3e4f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.594904] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.594904] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.594904] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1214.625899] env[62952]: DEBUG oslo_vmware.api [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367806, 'name': PowerOnVM_Task, 'duration_secs': 0.473991} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.626183] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1214.626389] env[62952]: INFO nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1214.626569] env[62952]: DEBUG nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1214.627339] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041d9c4f-c020-4dab-be16-00783f3e9bbb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.704256] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]520e8295-4c47-1241-79c6-018b20c3e4f8, 'name': SearchDatastore_Task, 'duration_secs': 0.012629} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.704556] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.704810] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1214.705081] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d7c898d-0e87-4393-beb0-26bb8b99dc11 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.711977] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1214.711977] env[62952]: value = "task-1367807" [ 1214.711977] env[62952]: _type = "Task" [ 1214.711977] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.719805] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.146266] env[62952]: INFO nova.compute.manager [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Took 11.36 seconds to build instance. [ 1215.221767] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367807, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444131} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.222044] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1215.222270] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1215.222520] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19570825-7d2b-457e-be2a-6f110ab405fb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.228852] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1215.228852] env[62952]: value = "task-1367808" [ 1215.228852] env[62952]: _type = "Task" [ 1215.228852] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.236820] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.649401] env[62952]: DEBUG oslo_concurrency.lockutils [None req-2d271fd4-6890-40b1-a879-6b65b09f936c tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.867s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.738355] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056026} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.739124] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1215.739484] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a83209-d638-4b00-b5f9-ea9afc8f7680 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.762351] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1215.762681] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3296d35f-9c38-4139-89d7-2f04873fa64e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.782850] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1215.782850] env[62952]: value = "task-1367809" [ 1215.782850] env[62952]: _type = "Task" [ 1215.782850] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.793115] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.813478] env[62952]: DEBUG nova.compute.manager [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Received event network-changed-199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1215.813732] env[62952]: DEBUG nova.compute.manager [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Refreshing instance network info cache due to event network-changed-199c0f83-9683-4568-806b-f01b051f2f0c. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1215.813987] env[62952]: DEBUG oslo_concurrency.lockutils [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] Acquiring lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1215.814200] env[62952]: DEBUG oslo_concurrency.lockutils [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] Acquired lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.814412] env[62952]: DEBUG nova.network.neutron [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Refreshing network info cache for port 199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1216.293066] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367809, 'name': ReconfigVM_Task, 'duration_secs': 0.272227} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.293521] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Reconfigured VM instance instance-0000006e to attach disk [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1216.294385] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b187c900-3806-4e3f-9713-b95dd1910caa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.300098] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1216.300098] env[62952]: value = "task-1367810" [ 1216.300098] env[62952]: _type = "Task" [ 1216.300098] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.307703] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367810, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.530766] env[62952]: DEBUG nova.network.neutron [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updated VIF entry in instance network info cache for port 199c0f83-9683-4568-806b-f01b051f2f0c. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1216.531161] env[62952]: DEBUG nova.network.neutron [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updating instance_info_cache with network_info: [{"id": "199c0f83-9683-4568-806b-f01b051f2f0c", "address": "fa:16:3e:75:e1:0d", "network": {"id": "ea2a1f8e-93dc-496e-bd9c-74ea26035ec3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1468645541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfc390d64c0463190f071f3f62936dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap199c0f83-96", "ovs_interfaceid": "199c0f83-9683-4568-806b-f01b051f2f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.634302] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.634495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.634677] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1216.811008] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367810, 'name': Rename_Task, 'duration_secs': 0.13306} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.811450] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1216.811716] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6a19ca7-ae76-4741-a1cc-feebffd4b864 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.818831] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1216.818831] env[62952]: value = "task-1367811" [ 1216.818831] env[62952]: _type = "Task" [ 1216.818831] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.830066] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.034311] env[62952]: DEBUG oslo_concurrency.lockutils [req-522b3351-b51d-4fd4-a5f3-b5ddebddc286 req-02e15b9b-1fe3-4e93-9b13-3625a3b2ec0b service nova] Releasing lock "refresh_cache-dd7d45f3-31bb-43c3-9317-2142d29b6ef5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.228207] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.228526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.228756] env[62952]: INFO nova.compute.manager [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Shelving [ 1217.329544] env[62952]: DEBUG oslo_vmware.api [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367811, 'name': PowerOnVM_Task, 'duration_secs': 0.436795} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.330046] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1217.330279] env[62952]: INFO nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1217.330467] env[62952]: DEBUG nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1217.331255] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d5303c-34f4-431b-8461-7f241a97329a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.736219] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1217.736499] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6806449b-5de4-45b3-bbb1-1860d5ddb1fe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.744120] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1217.744120] env[62952]: value = "task-1367812" [ 1217.744120] env[62952]: _type = "Task" [ 1217.744120] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.752016] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.849826] env[62952]: INFO nova.compute.manager [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Took 13.48 seconds to build instance. [ 1217.853417] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updating instance_info_cache with network_info: [{"id": "65c2ecad-019b-4edb-a82c-3cc676089eab", "address": "fa:16:3e:f5:06:67", "network": {"id": "6f239bd6-4ff6-4c0d-a685-cc7b700a1bbd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-294828919-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8c2a296e8fe4805b9bb761f91a407f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c2ecad-01", "ovs_interfaceid": "65c2ecad-019b-4edb-a82c-3cc676089eab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.253823] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367812, 'name': PowerOffVM_Task, 'duration_secs': 0.214859} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.254112] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1218.254885] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddc561f-59d2-409e-b244-31c8365a6133 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.273155] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488de733-d578-4b51-93e5-417996925844 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.352477] env[62952]: DEBUG oslo_concurrency.lockutils [None req-b1110fd9-3e5e-45da-83fd-01c0d6a8dab3 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.986s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.355040] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.355232] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1218.355421] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.355658] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.355831] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.355984] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.356167] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1218.783017] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1218.783373] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dcf61d72-1cfc-4cb9-8d79-802a610092e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.791337] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1218.791337] env[62952]: value = "task-1367813" [ 1218.791337] env[62952]: _type = "Task" [ 1218.791337] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.802024] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367813, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.816498] env[62952]: DEBUG nova.compute.manager [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1218.816498] env[62952]: DEBUG nova.compute.manager [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing instance network info cache due to event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1218.816498] env[62952]: DEBUG oslo_concurrency.lockutils [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.817563] env[62952]: DEBUG oslo_concurrency.lockutils [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.817563] env[62952]: DEBUG nova.network.neutron [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1219.302608] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367813, 'name': CreateSnapshot_Task, 'duration_secs': 0.396394} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.302918] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1219.303799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7141449e-1c76-4d7b-930b-3ce488584569 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.516473] env[62952]: DEBUG nova.network.neutron [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updated VIF entry in instance network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1219.516869] env[62952]: DEBUG nova.network.neutron [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.825462] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1219.825797] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cd79b292-50fe-48e1-b4b0-68ae1be9b92f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.834303] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1219.834303] env[62952]: value = "task-1367814" [ 1219.834303] env[62952]: _type = "Task" [ 1219.834303] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.842997] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367814, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.020647] env[62952]: DEBUG oslo_concurrency.lockutils [req-fda9bc5e-98f4-4a86-be46-ddbd066d7fd4 req-97a1cbc7-8a2f-4f68-bf75-69ddc5051460 service nova] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.344872] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367814, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.844678] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367814, 'name': CloneVM_Task} progress is 95%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.345300] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367814, 'name': CloneVM_Task, 'duration_secs': 1.142912} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.345679] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Created linked-clone VM from snapshot [ 1221.346275] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c150b94b-de8f-47e7-bfea-d61a4f22d13f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.353381] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Uploading image fd617927-f9a5-48f0-801c-2e6aeed6070f {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1221.378646] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1221.378646] env[62952]: value = "vm-291001" [ 1221.378646] env[62952]: _type = "VirtualMachine" [ 1221.378646] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1221.378941] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b628a17b-8aad-438f-b204-d754e392160b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.386159] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease: (returnval){ [ 1221.386159] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279ad3d-4683-abaa-7af0-328333bc1456" [ 1221.386159] env[62952]: _type = "HttpNfcLease" [ 1221.386159] env[62952]: } obtained for exporting VM: (result){ [ 1221.386159] env[62952]: value = "vm-291001" [ 1221.386159] env[62952]: _type = "VirtualMachine" [ 1221.386159] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1221.386680] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the lease: (returnval){ [ 1221.386680] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279ad3d-4683-abaa-7af0-328333bc1456" [ 1221.386680] env[62952]: _type = "HttpNfcLease" [ 1221.386680] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1221.392561] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1221.392561] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279ad3d-4683-abaa-7af0-328333bc1456" [ 1221.392561] env[62952]: _type = "HttpNfcLease" [ 1221.392561] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1221.894015] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1221.894015] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279ad3d-4683-abaa-7af0-328333bc1456" [ 1221.894015] env[62952]: _type = "HttpNfcLease" [ 1221.894015] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1221.894360] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1221.894360] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5279ad3d-4683-abaa-7af0-328333bc1456" [ 1221.894360] env[62952]: _type = "HttpNfcLease" [ 1221.894360] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1221.895091] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d795e7-1eea-4335-b777-70ffe8c355fc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.902147] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36146-6390-f850-ef8a-0d2d0270fa71/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1221.902322] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36146-6390-f850-ef8a-0d2d0270fa71/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1221.989044] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1e196574-7974-45e0-b710-8f7228541401 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.705011] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36146-6390-f850-ef8a-0d2d0270fa71/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1228.706040] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e705f8e7-9e05-45cc-b611-458ae2c6890d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.712047] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36146-6390-f850-ef8a-0d2d0270fa71/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1228.712225] env[62952]: ERROR oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36146-6390-f850-ef8a-0d2d0270fa71/disk-0.vmdk due to incomplete transfer. [ 1228.712454] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fb587087-dda8-4fe6-bb1b-75aa242a7935 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.718662] env[62952]: DEBUG oslo_vmware.rw_handles [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36146-6390-f850-ef8a-0d2d0270fa71/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1228.718820] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Uploaded image fd617927-f9a5-48f0-801c-2e6aeed6070f to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1228.721140] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1228.721362] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-486cfe40-e47b-4b44-9849-fa8faa60a78c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.726152] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1228.726152] env[62952]: value = "task-1367816" [ 1228.726152] env[62952]: _type = "Task" [ 1228.726152] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.733240] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367816, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.235635] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367816, 'name': Destroy_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.736565] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367816, 'name': Destroy_Task, 'duration_secs': 0.528812} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.736997] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Destroyed the VM [ 1229.737081] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1229.737307] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5a594273-7cd2-4d46-be7c-c2a72f642837 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.743557] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1229.743557] env[62952]: value = "task-1367817" [ 1229.743557] env[62952]: _type = "Task" [ 1229.743557] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.751425] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367817, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.253174] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367817, 'name': RemoveSnapshot_Task, 'duration_secs': 0.313493} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.253435] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1230.253719] env[62952]: DEBUG nova.compute.manager [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1230.254494] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a21a1e8-7caf-42c4-a827-6a789cb86dc4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.765987] env[62952]: INFO nova.compute.manager [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Shelve offloading [ 1230.768142] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1230.768142] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be357e53-14e7-4b46-82cd-2e767bd70d66 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.775442] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1230.775442] env[62952]: value = "task-1367818" [ 1230.775442] env[62952]: _type = "Task" [ 1230.775442] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.783082] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.287658] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1231.287943] env[62952]: DEBUG nova.compute.manager [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1231.288725] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620d8c29-8e70-4406-b5e1-59c222cd0f74 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.294361] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1231.294545] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.294733] env[62952]: DEBUG nova.network.neutron [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1231.996641] env[62952]: DEBUG nova.network.neutron [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.499182] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.766237] env[62952]: DEBUG nova.compute.manager [req-5cee6644-476e-4367-bf99-5bf65dc5dde2 req-e808e3ce-927c-4950-b4cd-39af8a94ec2d service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-vif-unplugged-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1232.766489] env[62952]: DEBUG oslo_concurrency.lockutils [req-5cee6644-476e-4367-bf99-5bf65dc5dde2 req-e808e3ce-927c-4950-b4cd-39af8a94ec2d service nova] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.766714] env[62952]: DEBUG oslo_concurrency.lockutils [req-5cee6644-476e-4367-bf99-5bf65dc5dde2 req-e808e3ce-927c-4950-b4cd-39af8a94ec2d service nova] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.766900] env[62952]: DEBUG oslo_concurrency.lockutils [req-5cee6644-476e-4367-bf99-5bf65dc5dde2 req-e808e3ce-927c-4950-b4cd-39af8a94ec2d service nova] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.767090] env[62952]: DEBUG nova.compute.manager [req-5cee6644-476e-4367-bf99-5bf65dc5dde2 req-e808e3ce-927c-4950-b4cd-39af8a94ec2d service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] No waiting events found dispatching network-vif-unplugged-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1232.767222] env[62952]: WARNING nova.compute.manager [req-5cee6644-476e-4367-bf99-5bf65dc5dde2 req-e808e3ce-927c-4950-b4cd-39af8a94ec2d service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received unexpected event network-vif-unplugged-551a3758-07da-4c7a-acda-583eb14f995f for instance with vm_state shelved and task_state shelving_offloading. [ 1232.827818] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1232.828844] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d99b14-0aac-4469-b188-00b42c97a8aa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.837407] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1232.837607] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b30e43d7-dd64-4acf-9e8e-e283130e6746 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.906066] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1232.906305] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1232.906515] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleting the datastore file [datastore2] 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1232.906785] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b09245c8-0c88-4c2f-b30f-252fa7e0928c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.913303] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1232.913303] env[62952]: value = "task-1367820" [ 1232.913303] env[62952]: _type = "Task" [ 1232.913303] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.920505] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367820, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.422843] env[62952]: DEBUG oslo_vmware.api [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367820, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238178} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.423127] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1233.423500] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1233.423500] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1233.463272] env[62952]: INFO nova.scheduler.client.report [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted allocations for instance 27e78e78-3a7f-4ad4-b939-438cb2b90b55 [ 1233.966574] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.966860] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.967104] env[62952]: DEBUG nova.objects.instance [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'resources' on Instance uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.469849] env[62952]: DEBUG nova.objects.instance [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'numa_topology' on Instance uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.770804] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.771908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.771908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.771908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.771908] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.774014] env[62952]: INFO nova.compute.manager [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Terminating instance [ 1234.776076] env[62952]: DEBUG nova.compute.manager [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1234.776299] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1234.776589] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2369f6bc-d319-4a59-9cc9-df5b344d90ed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.784282] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1234.784282] env[62952]: value = "task-1367821" [ 1234.784282] env[62952]: _type = "Task" [ 1234.784282] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.791502] env[62952]: DEBUG nova.compute.manager [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-changed-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1234.791711] env[62952]: DEBUG nova.compute.manager [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing instance network info cache due to event network-changed-551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1234.792054] env[62952]: DEBUG oslo_concurrency.lockutils [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.792167] env[62952]: DEBUG oslo_concurrency.lockutils [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.792382] env[62952]: DEBUG nova.network.neutron [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1234.796883] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.973688] env[62952]: DEBUG nova.objects.base [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Object Instance<27e78e78-3a7f-4ad4-b939-438cb2b90b55> lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1234.991557] env[62952]: DEBUG nova.scheduler.client.report [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Refreshing inventories for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1235.005451] env[62952]: DEBUG nova.scheduler.client.report [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Updating ProviderTree inventory for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1235.005684] env[62952]: DEBUG nova.compute.provider_tree [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Updating inventory in ProviderTree for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1235.016364] env[62952]: DEBUG nova.scheduler.client.report [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Refreshing aggregate associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, aggregates: None {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1235.033274] env[62952]: DEBUG nova.scheduler.client.report [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Refreshing trait associations for resource provider 9b9257d4-3be7-42a5-ae8c-67cf64823681, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=62952) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1235.100277] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b213afc3-bf5b-4b35-b933-883d1e19b9d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.107417] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2aaeded-9874-44d6-b82c-f3ebad64eaa3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.136125] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eb7226-658a-4550-8f75-3fd9c1bc246c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.143073] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cd6159-a7b1-46aa-8e88-8f1495d8782c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.155769] env[62952]: DEBUG nova.compute.provider_tree [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.295139] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367821, 'name': PowerOffVM_Task, 'duration_secs': 0.190504} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.295352] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1235.295552] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1235.295746] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290989', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'name': 'volume-188d94f2-707a-48de-b891-ba671ade4185', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '6606e84c-36cd-4519-beff-096a8f1bc09e', 'attached_at': '2024-10-26T01:44:03.000000', 'detached_at': '', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'serial': '188d94f2-707a-48de-b891-ba671ade4185'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1235.296534] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4ec6b4-b5fc-4889-a7f9-1e2b07422533 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.315682] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb59f0b-7a29-457c-92ff-68a721bb5c71 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.323466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da548d64-1b73-4f7e-9118-16b905c5a26b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.341242] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96aa9cac-4308-4c51-b7d2-e64a882879b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.355935] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] The volume has not been displaced from its original location: [datastore2] volume-188d94f2-707a-48de-b891-ba671ade4185/volume-188d94f2-707a-48de-b891-ba671ade4185.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1235.361169] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1235.361507] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b82c523-1918-43c6-a173-dfdcdb946272 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.379470] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1235.379470] env[62952]: value = "task-1367822" [ 1235.379470] env[62952]: _type = "Task" [ 1235.379470] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.387367] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367822, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.564601] env[62952]: DEBUG nova.network.neutron [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updated VIF entry in instance network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1235.564980] env[62952]: DEBUG nova.network.neutron [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap551a3758-07", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.659110] env[62952]: DEBUG nova.scheduler.client.report [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1235.889258] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367822, 'name': ReconfigVM_Task, 'duration_secs': 0.146356} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.889633] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1235.894220] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7f34810-ff63-4d16-8c12-459b3e336bc1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.909341] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1235.909341] env[62952]: value = "task-1367823" [ 1235.909341] env[62952]: _type = "Task" [ 1235.909341] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.918803] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367823, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.069169] env[62952]: DEBUG oslo_concurrency.lockutils [req-b972fe38-6a80-408a-9d34-d1ef926c2782 req-066720eb-df69-4094-ab0d-000604f53eb5 service nova] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1236.083059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.163847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.197s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.418916] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367823, 'name': ReconfigVM_Task, 'duration_secs': 0.205793} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.419218] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290989', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'name': 'volume-188d94f2-707a-48de-b891-ba671ade4185', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '6606e84c-36cd-4519-beff-096a8f1bc09e', 'attached_at': '2024-10-26T01:44:03.000000', 'detached_at': '', 'volume_id': '188d94f2-707a-48de-b891-ba671ade4185', 'serial': '188d94f2-707a-48de-b891-ba671ade4185'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1236.419498] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1236.420236] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729238ab-790a-4539-ae94-1993f1215c35 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.426154] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1236.426404] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d173c0a1-afcb-48aa-be44-a2e89e21f19f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.487973] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1236.488225] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1236.488404] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore2] 6606e84c-36cd-4519-beff-096a8f1bc09e {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1236.488741] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6c4fc72-f4bc-43d5-8892-46c0f70e0d15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.494946] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1236.494946] env[62952]: value = "task-1367825" [ 1236.494946] env[62952]: _type = "Task" [ 1236.494946] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.502456] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.673567] env[62952]: DEBUG oslo_concurrency.lockutils [None req-42626262-b75d-42f8-a7cd-29b3ae93ea5c tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.445s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.674892] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.592s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.675041] env[62952]: INFO nova.compute.manager [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Unshelving [ 1237.005520] env[62952]: DEBUG oslo_vmware.api [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367825, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.078493} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.005737] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1237.005927] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1237.006126] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1237.006307] env[62952]: INFO nova.compute.manager [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1237.006579] env[62952]: DEBUG oslo.service.loopingcall [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1237.006773] env[62952]: DEBUG nova.compute.manager [-] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1237.006867] env[62952]: DEBUG nova.network.neutron [-] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1237.411530] env[62952]: DEBUG nova.compute.manager [req-6a6bbccc-7638-474f-ad8b-a6a200e9a698 req-cbfcf48c-f337-441c-b84b-d3a1f1b5c5e9 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Received event network-vif-deleted-de8307b9-6263-4138-befe-40f603f1bea4 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1237.411773] env[62952]: INFO nova.compute.manager [req-6a6bbccc-7638-474f-ad8b-a6a200e9a698 req-cbfcf48c-f337-441c-b84b-d3a1f1b5c5e9 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Neutron deleted interface de8307b9-6263-4138-befe-40f603f1bea4; detaching it from the instance and deleting it from the info cache [ 1237.411951] env[62952]: DEBUG nova.network.neutron [req-6a6bbccc-7638-474f-ad8b-a6a200e9a698 req-cbfcf48c-f337-441c-b84b-d3a1f1b5c5e9 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.684546] env[62952]: DEBUG nova.compute.utils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1237.888037] env[62952]: DEBUG nova.network.neutron [-] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.914569] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20243cd6-0450-4f44-82c4-ab546617b269 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.924770] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2271b3-2232-46f7-9a26-c3b550498a03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.950669] env[62952]: DEBUG nova.compute.manager [req-6a6bbccc-7638-474f-ad8b-a6a200e9a698 req-cbfcf48c-f337-441c-b84b-d3a1f1b5c5e9 service nova] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Detach interface failed, port_id=de8307b9-6263-4138-befe-40f603f1bea4, reason: Instance 6606e84c-36cd-4519-beff-096a8f1bc09e could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1238.188652] env[62952]: INFO nova.virt.block_device [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Booting with volume ca518c35-2fc1-4cac-b6d5-c2c01d5b057e at /dev/sdb [ 1238.224029] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0beabb44-a385-4a6c-a891-8ed576c7997e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.232948] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c1e78a-b288-42bf-95ba-32701a23f81a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.259242] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f97d130e-b02f-4435-9b93-941724352193 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.266717] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5919ec-5493-4407-b85e-56a9bdfcf490 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.290737] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879b20ba-cb12-4691-848a-7a0aaddb13a2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.296754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea01219c-cb56-41de-8383-1c9dc8a5cadc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.309137] env[62952]: DEBUG nova.virt.block_device [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating existing volume attachment record: d7f573f2-3630-4b2c-ac2d-b8b672bd4e43 {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1238.390614] env[62952]: INFO nova.compute.manager [-] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Took 1.38 seconds to deallocate network for instance. [ 1238.932175] env[62952]: INFO nova.compute.manager [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Took 0.54 seconds to detach 1 volumes for instance. [ 1238.934353] env[62952]: DEBUG nova.compute.manager [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Deleting volume: 188d94f2-707a-48de-b891-ba671ade4185 {{(pid=62952) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1239.469896] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.470189] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.470413] env[62952]: DEBUG nova.objects.instance [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'resources' on Instance uuid 6606e84c-36cd-4519-beff-096a8f1bc09e {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.047466] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d496e05b-3687-446b-a189-20f2c3de7559 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.056376] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4063fda-79df-401f-beea-b76541c79a8c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.087764] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f75a12-c837-4ced-bf6b-1f599c9cee24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.094256] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f638df5e-1db0-46d9-9f2a-ca9f4cc6d14c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.106671] env[62952]: DEBUG nova.compute.provider_tree [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1240.609534] env[62952]: DEBUG nova.scheduler.client.report [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1241.114712] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.644s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.131726] env[62952]: INFO nova.scheduler.client.report [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted allocations for instance 6606e84c-36cd-4519-beff-096a8f1bc09e [ 1241.639264] env[62952]: DEBUG oslo_concurrency.lockutils [None req-9fbe080b-1486-44c1-9541-d08be1031d23 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "6606e84c-36cd-4519-beff-096a8f1bc09e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.868s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.444357] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.444732] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.444859] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.445060] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.445240] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.447460] env[62952]: INFO nova.compute.manager [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Terminating instance [ 1242.449198] env[62952]: DEBUG nova.compute.manager [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1242.449398] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1242.450238] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1160df-a9ba-4412-98f5-744b1a233abd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.458372] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1242.458633] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09d765a3-947b-4d7d-a0c8-b5084af9ad79 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.465265] env[62952]: DEBUG oslo_vmware.api [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1242.465265] env[62952]: value = "task-1367831" [ 1242.465265] env[62952]: _type = "Task" [ 1242.465265] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.473045] env[62952]: DEBUG oslo_vmware.api [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.976245] env[62952]: DEBUG oslo_vmware.api [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367831, 'name': PowerOffVM_Task, 'duration_secs': 0.182142} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.976584] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1242.976708] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1242.976962] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bcc7199-2146-4a30-86db-597d5f1e5612 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.042529] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1243.042742] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1243.042918] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore1] 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1243.043326] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94c53cba-1e1d-4a69-a36e-8f89cf08d607 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.049773] env[62952]: DEBUG oslo_vmware.api [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1243.049773] env[62952]: value = "task-1367833" [ 1243.049773] env[62952]: _type = "Task" [ 1243.049773] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.057706] env[62952]: DEBUG oslo_vmware.api [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.559716] env[62952]: DEBUG oslo_vmware.api [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140088} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.560130] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1243.560180] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1243.560357] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1243.560552] env[62952]: INFO nova.compute.manager [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1243.560790] env[62952]: DEBUG oslo.service.loopingcall [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1243.560977] env[62952]: DEBUG nova.compute.manager [-] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1243.561082] env[62952]: DEBUG nova.network.neutron [-] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1243.917562] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.917844] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.918217] env[62952]: DEBUG nova.objects.instance [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'pci_requests' on Instance uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.056650] env[62952]: DEBUG nova.compute.manager [req-69cf02e9-8ddf-42ef-9d03-6d3f2fdeaa02 req-890e2eb3-6076-48cd-b583-6f0d8a399c24 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Received event network-vif-deleted-65c2ecad-019b-4edb-a82c-3cc676089eab {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1244.056838] env[62952]: INFO nova.compute.manager [req-69cf02e9-8ddf-42ef-9d03-6d3f2fdeaa02 req-890e2eb3-6076-48cd-b583-6f0d8a399c24 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Neutron deleted interface 65c2ecad-019b-4edb-a82c-3cc676089eab; detaching it from the instance and deleting it from the info cache [ 1244.056998] env[62952]: DEBUG nova.network.neutron [req-69cf02e9-8ddf-42ef-9d03-6d3f2fdeaa02 req-890e2eb3-6076-48cd-b583-6f0d8a399c24 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.422688] env[62952]: DEBUG nova.objects.instance [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'numa_topology' on Instance uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.534220] env[62952]: DEBUG nova.network.neutron [-] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.560061] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b896c38b-2cc8-4c26-ba9c-406953566217 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.569292] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433208f8-43d5-4dd9-8f1e-6d9507493661 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.593087] env[62952]: DEBUG nova.compute.manager [req-69cf02e9-8ddf-42ef-9d03-6d3f2fdeaa02 req-890e2eb3-6076-48cd-b583-6f0d8a399c24 service nova] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Detach interface failed, port_id=65c2ecad-019b-4edb-a82c-3cc676089eab, reason: Instance 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1244.925445] env[62952]: INFO nova.compute.claims [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1245.037035] env[62952]: INFO nova.compute.manager [-] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Took 1.48 seconds to deallocate network for instance. [ 1245.543547] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.003642] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38312ab1-283a-4f12-b467-7076a8109edf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.011192] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056e077f-77aa-4157-9f00-4cebe9b2d0ab {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.040625] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b2c695-459e-4965-a24b-9160d3befbd9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.047142] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c7a2af-c643-4604-ba01-d3930b4a9fe5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.059475] env[62952]: DEBUG nova.compute.provider_tree [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1246.562769] env[62952]: DEBUG nova.scheduler.client.report [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1247.067682] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.150s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.069900] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.526s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.070166] env[62952]: DEBUG nova.objects.instance [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'resources' on Instance uuid 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.096592] env[62952]: INFO nova.network.neutron [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating port 551a3758-07da-4c7a-acda-583eb14f995f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1247.638688] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a83450e-99ac-46a0-a078-2af25fc58f07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.645950] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f68a9aa-1426-4c66-91dc-b348eb5e5cc0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.675694] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b464b2-dab5-464a-9ad6-bb610b63ff3f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.682328] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e4d4b9-6e1a-4399-9d53-b1d6ab7829d2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.694876] env[62952]: DEBUG nova.compute.provider_tree [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.198098] env[62952]: DEBUG nova.scheduler.client.report [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1248.539563] env[62952]: DEBUG nova.compute.manager [req-473b4c90-4f80-4b40-a9b5-35c91be0a1f8 req-ede8e064-971d-4fd9-b160-e5bb0b980e0e service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-vif-plugged-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1248.539788] env[62952]: DEBUG oslo_concurrency.lockutils [req-473b4c90-4f80-4b40-a9b5-35c91be0a1f8 req-ede8e064-971d-4fd9-b160-e5bb0b980e0e service nova] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.540015] env[62952]: DEBUG oslo_concurrency.lockutils [req-473b4c90-4f80-4b40-a9b5-35c91be0a1f8 req-ede8e064-971d-4fd9-b160-e5bb0b980e0e service nova] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.540251] env[62952]: DEBUG oslo_concurrency.lockutils [req-473b4c90-4f80-4b40-a9b5-35c91be0a1f8 req-ede8e064-971d-4fd9-b160-e5bb0b980e0e service nova] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.540360] env[62952]: DEBUG nova.compute.manager [req-473b4c90-4f80-4b40-a9b5-35c91be0a1f8 req-ede8e064-971d-4fd9-b160-e5bb0b980e0e service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] No waiting events found dispatching network-vif-plugged-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1248.540514] env[62952]: WARNING nova.compute.manager [req-473b4c90-4f80-4b40-a9b5-35c91be0a1f8 req-ede8e064-971d-4fd9-b160-e5bb0b980e0e service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received unexpected event network-vif-plugged-551a3758-07da-4c7a-acda-583eb14f995f for instance with vm_state shelved_offloaded and task_state spawning. [ 1248.622165] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.622326] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.622507] env[62952]: DEBUG nova.network.neutron [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1248.704063] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.722191] env[62952]: INFO nova.scheduler.client.report [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted allocations for instance 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5 [ 1249.230218] env[62952]: DEBUG oslo_concurrency.lockutils [None req-c22cc15b-9264-4867-8eec-44dc5a2ff4a1 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "3cc1f262-3e94-4a7e-997c-fbd24bad0aa5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.785s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.327444] env[62952]: DEBUG nova.network.neutron [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.829792] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1249.856012] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3f46e9eff9de8ba8b977544a9e93db1b',container_format='bare',created_at=2024-10-26T01:44:23Z,direct_url=,disk_format='vmdk',id=fd617927-f9a5-48f0-801c-2e6aeed6070f,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1950053080-shelved',owner='a2c2a97031df4b75b835a96b41793c74',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-26T01:44:35Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1249.856365] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1249.856544] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1249.856732] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1249.856882] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1249.857052] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1249.857274] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1249.857786] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1249.857786] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1249.857942] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1249.858022] env[62952]: DEBUG nova.virt.hardware [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1249.858884] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95cdd86-798f-4ac2-95f4-f7a724e783ff {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.867182] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64663bad-0087-4e0f-8196-8915c95e74f3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.881437] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e2:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd22cb4ec-277f-41ee-8aba-b3d54442b93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '551a3758-07da-4c7a-acda-583eb14f995f', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1249.888867] env[62952]: DEBUG oslo.service.loopingcall [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1249.889135] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1249.889370] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3fad3058-7003-418e-a14a-5f4b865e1ced {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.908144] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1249.908144] env[62952]: value = "task-1367835" [ 1249.908144] env[62952]: _type = "Task" [ 1249.908144] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.915495] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367835, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.417787] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367835, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.570705] env[62952]: DEBUG nova.compute.manager [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-changed-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1250.570910] env[62952]: DEBUG nova.compute.manager [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing instance network info cache due to event network-changed-551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1250.571135] env[62952]: DEBUG oslo_concurrency.lockutils [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] Acquiring lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.571284] env[62952]: DEBUG oslo_concurrency.lockutils [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] Acquired lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.571450] env[62952]: DEBUG nova.network.neutron [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Refreshing network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1250.919016] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367835, 'name': CreateVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.285168] env[62952]: DEBUG nova.network.neutron [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updated VIF entry in instance network info cache for port 551a3758-07da-4c7a-acda-583eb14f995f. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1251.285544] env[62952]: DEBUG nova.network.neutron [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [{"id": "551a3758-07da-4c7a-acda-583eb14f995f", "address": "fa:16:3e:73:e2:2d", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap551a3758-07", "ovs_interfaceid": "551a3758-07da-4c7a-acda-583eb14f995f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.419239] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367835, 'name': CreateVM_Task, 'duration_secs': 1.340173} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.419578] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1251.419995] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.420187] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.420576] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1251.420823] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aa2617e-ca90-49a3-9a06-64810eabaff5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.425263] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1251.425263] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5276ad38-00cb-c026-daab-76b2cda22f04" [ 1251.425263] env[62952]: _type = "Task" [ 1251.425263] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.433734] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5276ad38-00cb-c026-daab-76b2cda22f04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.788266] env[62952]: DEBUG oslo_concurrency.lockutils [req-b10db5fa-ea39-4254-8c02-968d9db54513 req-52ab29fc-ff69-4212-aa03-e94fc722232c service nova] Releasing lock "refresh_cache-27e78e78-3a7f-4ad4-b939-438cb2b90b55" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.935707] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.935962] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Processing image fd617927-f9a5-48f0-801c-2e6aeed6070f {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1251.936248] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.936452] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.936660] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1251.936927] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d821f34-aa5a-405a-9d35-e8ca2b5a68d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.944620] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1251.944785] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1251.945457] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb0cab52-e4f4-4fdc-a84b-8f98077cd46d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.950013] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1251.950013] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52e6f532-174d-67cb-2b6e-b91d28842f2c" [ 1251.950013] env[62952]: _type = "Task" [ 1251.950013] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.957027] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52e6f532-174d-67cb-2b6e-b91d28842f2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.462516] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1252.462860] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Fetch image to [datastore1] OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175/OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1252.463092] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Downloading stream optimized image fd617927-f9a5-48f0-801c-2e6aeed6070f to [datastore1] OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175/OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175.vmdk on the data store datastore1 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1252.463353] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Downloading image file data fd617927-f9a5-48f0-801c-2e6aeed6070f to the ESX as VM named 'OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1252.543217] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1252.543217] env[62952]: value = "resgroup-9" [ 1252.543217] env[62952]: _type = "ResourcePool" [ 1252.543217] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1252.543533] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9a36342f-e0de-44f8-8d7b-2b7d3c500729 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.564734] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease: (returnval){ [ 1252.564734] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524fc113-c442-fb36-a259-545196482c67" [ 1252.564734] env[62952]: _type = "HttpNfcLease" [ 1252.564734] env[62952]: } obtained for vApp import into resource pool (val){ [ 1252.564734] env[62952]: value = "resgroup-9" [ 1252.564734] env[62952]: _type = "ResourcePool" [ 1252.564734] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1252.565030] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the lease: (returnval){ [ 1252.565030] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524fc113-c442-fb36-a259-545196482c67" [ 1252.565030] env[62952]: _type = "HttpNfcLease" [ 1252.565030] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1252.570657] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1252.570657] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524fc113-c442-fb36-a259-545196482c67" [ 1252.570657] env[62952]: _type = "HttpNfcLease" [ 1252.570657] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1253.073057] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1253.073057] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524fc113-c442-fb36-a259-545196482c67" [ 1253.073057] env[62952]: _type = "HttpNfcLease" [ 1253.073057] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1253.073367] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1253.073367] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]524fc113-c442-fb36-a259-545196482c67" [ 1253.073367] env[62952]: _type = "HttpNfcLease" [ 1253.073367] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1253.074092] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545944be-fade-4b0c-9c12-3e67e12c4d43 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.081949] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523afe71-558b-1584-0e37-f89cfd13dded/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1253.082145] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523afe71-558b-1584-0e37-f89cfd13dded/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1253.144892] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-26503b5b-14ed-46f3-a147-096a270c56e7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.157148] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "aef28168-98a7-4f65-80e7-731633339abf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.157463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "aef28168-98a7-4f65-80e7-731633339abf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.157730] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "aef28168-98a7-4f65-80e7-731633339abf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.157965] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "aef28168-98a7-4f65-80e7-731633339abf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.158199] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "aef28168-98a7-4f65-80e7-731633339abf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.160546] env[62952]: INFO nova.compute.manager [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Terminating instance [ 1253.162537] env[62952]: DEBUG nova.compute.manager [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1253.162766] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1253.163671] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc62f104-a045-41ec-bc61-f3c8a4d02afe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.171214] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1253.171494] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-358dc7ac-90ad-465c-a885-c529f832cf8a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.179736] env[62952]: DEBUG oslo_vmware.api [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1253.179736] env[62952]: value = "task-1367837" [ 1253.179736] env[62952]: _type = "Task" [ 1253.179736] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.188585] env[62952]: DEBUG oslo_vmware.api [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367837, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.350675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.350915] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.690744] env[62952]: DEBUG oslo_vmware.api [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367837, 'name': PowerOffVM_Task, 'duration_secs': 0.203708} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.691287] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1253.691287] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1253.691526] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14010ea7-9032-4e97-99e1-a9f3451263cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.854721] env[62952]: DEBUG nova.compute.utils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1253.859617] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1253.859867] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1253.860165] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleting the datastore file [datastore2] aef28168-98a7-4f65-80e7-731633339abf {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1253.860445] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3629726d-d127-48f1-b25b-97c3879d62e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.873459] env[62952]: DEBUG oslo_vmware.api [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for the task: (returnval){ [ 1253.873459] env[62952]: value = "task-1367839" [ 1253.873459] env[62952]: _type = "Task" [ 1253.873459] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.882700] env[62952]: DEBUG oslo_vmware.api [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.223541] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.223938] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.224153] env[62952]: INFO nova.compute.manager [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Shelving [ 1254.300411] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1254.300743] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523afe71-558b-1584-0e37-f89cfd13dded/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1254.304494] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dc9373-3780-4346-b0b3-bee30841608b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.311938] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523afe71-558b-1584-0e37-f89cfd13dded/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1254.312124] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523afe71-558b-1584-0e37-f89cfd13dded/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1254.312395] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-038ab882-125e-4b7a-9b06-ae854d1618d3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.361079] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.382941] env[62952]: DEBUG oslo_vmware.api [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Task: {'id': task-1367839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22696} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.383204] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1254.383391] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1254.383565] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1254.383737] env[62952]: INFO nova.compute.manager [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] [instance: aef28168-98a7-4f65-80e7-731633339abf] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1254.383977] env[62952]: DEBUG oslo.service.loopingcall [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1254.384183] env[62952]: DEBUG nova.compute.manager [-] [instance: aef28168-98a7-4f65-80e7-731633339abf] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1254.384280] env[62952]: DEBUG nova.network.neutron [-] [instance: aef28168-98a7-4f65-80e7-731633339abf] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1254.553219] env[62952]: DEBUG oslo_vmware.rw_handles [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523afe71-558b-1584-0e37-f89cfd13dded/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1254.553452] env[62952]: INFO nova.virt.vmwareapi.images [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Downloaded image file data fd617927-f9a5-48f0-801c-2e6aeed6070f [ 1254.554356] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4aec672-743f-4502-ad7f-444fdb48918e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.569635] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a7bf6bf-3b66-402d-84e6-ea4756f93fbc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.595386] env[62952]: INFO nova.virt.vmwareapi.images [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] The imported VM was unregistered [ 1254.598266] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1254.598426] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating directory with path [datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1254.599917] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f692da1-cdfb-4cd6-bede-1899f3ac8d56 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.610507] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created directory with path [datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1254.610706] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175/OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175.vmdk to [datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1254.611558] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c0a58228-5100-49e3-9f80-d37db7057d1f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.618750] env[62952]: DEBUG nova.compute.manager [req-540c4a3f-0c72-4419-9ba3-7cf90aed0f3d req-a8b6582b-d04c-4c99-bb17-3444420e2763 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Received event network-vif-deleted-f80edb6b-5326-4531-a02f-366484e828ce {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1254.618955] env[62952]: INFO nova.compute.manager [req-540c4a3f-0c72-4419-9ba3-7cf90aed0f3d req-a8b6582b-d04c-4c99-bb17-3444420e2763 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Neutron deleted interface f80edb6b-5326-4531-a02f-366484e828ce; detaching it from the instance and deleting it from the info cache [ 1254.619144] env[62952]: DEBUG nova.network.neutron [req-540c4a3f-0c72-4419-9ba3-7cf90aed0f3d req-a8b6582b-d04c-4c99-bb17-3444420e2763 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1254.621611] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1254.621611] env[62952]: value = "task-1367841" [ 1254.621611] env[62952]: _type = "Task" [ 1254.621611] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.630797] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367841, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.731387] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1254.731765] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db5210fd-a42f-4870-af43-bcd32481b3d8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.738148] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1254.738148] env[62952]: value = "task-1367842" [ 1254.738148] env[62952]: _type = "Task" [ 1254.738148] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.746013] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.103250] env[62952]: DEBUG nova.network.neutron [-] [instance: aef28168-98a7-4f65-80e7-731633339abf] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.123454] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a21f7eb-dbc7-4e4c-9a00-7710c8726d4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.135512] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367841, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.139362] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffec3cb0-cb15-49c2-b674-2dd59d8fbef5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.164882] env[62952]: DEBUG nova.compute.manager [req-540c4a3f-0c72-4419-9ba3-7cf90aed0f3d req-a8b6582b-d04c-4c99-bb17-3444420e2763 service nova] [instance: aef28168-98a7-4f65-80e7-731633339abf] Detach interface failed, port_id=f80edb6b-5326-4531-a02f-366484e828ce, reason: Instance aef28168-98a7-4f65-80e7-731633339abf could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1255.247613] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367842, 'name': PowerOffVM_Task, 'duration_secs': 0.234962} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.248302] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1255.249127] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9bd3d04-493a-4f41-8fb2-755d6e03b833 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.267219] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2189a620-43b5-400f-b2e5-6170953393a8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.433122] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.433481] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.433741] env[62952]: INFO nova.compute.manager [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Attaching volume 2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78 to /dev/sdb [ 1255.469645] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366cbfa9-c9d7-48f6-a6a9-8d98f5ef8bbe {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.477757] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b7d52c-a97f-4f6c-840c-12cb5f598095 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.491690] env[62952]: DEBUG nova.virt.block_device [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updating existing volume attachment record: 5fb46654-2f6f-4031-8e88-9dd4207a9f6d {{(pid=62952) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1255.606570] env[62952]: INFO nova.compute.manager [-] [instance: aef28168-98a7-4f65-80e7-731633339abf] Took 1.22 seconds to deallocate network for instance. [ 1255.633270] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367841, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.778830] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1255.779300] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d4c8fa71-ad80-4aa4-8784-38c5bb392d98 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.787782] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1255.787782] env[62952]: value = "task-1367844" [ 1255.787782] env[62952]: _type = "Task" [ 1255.787782] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.796692] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367844, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.114098] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.114453] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.114628] env[62952]: DEBUG nova.objects.instance [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lazy-loading 'resources' on Instance uuid aef28168-98a7-4f65-80e7-731633339abf {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.134323] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367841, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.297576] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367844, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.636050] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367841, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.692211] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd6ccce-3986-4ea5-ab48-9b252421bd2f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.699828] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02ab015-b223-44b5-aa68-dd8c22a14c8e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.730179] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52311b5e-3985-48d0-a14f-3f7f2c911d02 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.738107] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fac211-d36b-420c-88ca-c35989f2d55b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.752209] env[62952]: DEBUG nova.compute.provider_tree [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1256.799432] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367844, 'name': CreateSnapshot_Task, 'duration_secs': 0.867791} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.799831] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1256.800518] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16131ef9-7390-484c-aec5-fbc3de2d2f36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.135621] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367841, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.427221} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.135909] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175/OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175.vmdk to [datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk. [ 1257.136103] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Cleaning up location [datastore1] OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1257.136278] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_1a03b924-dc4d-48c9-b903-112de2496175 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1257.136516] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bbace0cf-edaf-433e-8a01-921423ff3d32 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.142775] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1257.142775] env[62952]: value = "task-1367845" [ 1257.142775] env[62952]: _type = "Task" [ 1257.142775] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.149884] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367845, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.256184] env[62952]: DEBUG nova.scheduler.client.report [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1257.318788] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1257.319180] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e83af31c-f19c-4be8-a844-f492d8c3b233 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.329152] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1257.329152] env[62952]: value = "task-1367846" [ 1257.329152] env[62952]: _type = "Task" [ 1257.329152] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.336476] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367846, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.653229] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367845, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035153} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.653513] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1257.653644] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.653923] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk to [datastore1] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1257.654216] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbf82ba8-16f9-4670-832a-17e761b89c2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.660934] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1257.660934] env[62952]: value = "task-1367847" [ 1257.660934] env[62952]: _type = "Task" [ 1257.660934] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.668295] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.761461] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.783094] env[62952]: INFO nova.scheduler.client.report [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Deleted allocations for instance aef28168-98a7-4f65-80e7-731633339abf [ 1257.840447] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367846, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.170938] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.291275] env[62952]: DEBUG oslo_concurrency.lockutils [None req-3a94969b-fbde-4914-8ae2-b320c5e4bb31 tempest-ServerActionsTestOtherA-1315771520 tempest-ServerActionsTestOtherA-1315771520-project-member] Lock "aef28168-98a7-4f65-80e7-731633339abf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.134s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1258.340205] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367846, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.672510] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.840635] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367846, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.173429] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.340431] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367846, 'name': CloneVM_Task, 'duration_secs': 1.929953} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.340724] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Created linked-clone VM from snapshot [ 1259.341538] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08f1a02-5b0d-4b31-9235-f6b4d9b5e743 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.349319] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Uploading image 2907e76b-65cb-4cbe-8115-6d1d708e65d3 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1259.379230] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1259.379230] env[62952]: value = "vm-291008" [ 1259.379230] env[62952]: _type = "VirtualMachine" [ 1259.379230] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1259.379542] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d328b607-2f51-4c65-98b4-8a9cef918510 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.387875] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease: (returnval){ [ 1259.387875] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52777b97-384e-8af5-2cd5-a082473a3086" [ 1259.387875] env[62952]: _type = "HttpNfcLease" [ 1259.387875] env[62952]: } obtained for exporting VM: (result){ [ 1259.387875] env[62952]: value = "vm-291008" [ 1259.387875] env[62952]: _type = "VirtualMachine" [ 1259.387875] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1259.388360] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the lease: (returnval){ [ 1259.388360] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52777b97-384e-8af5-2cd5-a082473a3086" [ 1259.388360] env[62952]: _type = "HttpNfcLease" [ 1259.388360] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1259.395387] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1259.395387] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52777b97-384e-8af5-2cd5-a082473a3086" [ 1259.395387] env[62952]: _type = "HttpNfcLease" [ 1259.395387] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1259.674297] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.896734] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1259.896734] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52777b97-384e-8af5-2cd5-a082473a3086" [ 1259.896734] env[62952]: _type = "HttpNfcLease" [ 1259.896734] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1259.897226] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1259.897226] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52777b97-384e-8af5-2cd5-a082473a3086" [ 1259.897226] env[62952]: _type = "HttpNfcLease" [ 1259.897226] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1259.898031] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1a5419-cbbb-48a8-88fb-7756525e6d03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.905932] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5240b2b9-3279-c687-f1c6-c8d72982ddb4/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1259.906202] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5240b2b9-3279-c687-f1c6-c8d72982ddb4/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1260.014045] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d6d532ab-d81b-42f6-aa7d-cf1a4161fd05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.038441] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1260.038689] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291006', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'name': 'volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd7d45f3-31bb-43c3-9317-2142d29b6ef5', 'attached_at': '', 'detached_at': '', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'serial': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1260.039824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b827b8a-1da5-42ff-81eb-4852cb407e0c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.062407] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e871c53-3425-4824-8bd4-7ec61bbc8a3a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.089163] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78/volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.089521] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d8e1d25-2346-4889-83ca-c1a4e3526977 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.108898] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1260.108898] env[62952]: value = "task-1367850" [ 1260.108898] env[62952]: _type = "Task" [ 1260.108898] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.121315] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367850, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.175214] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.618848] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367850, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.675275] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367847, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.729112} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.675670] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fd617927-f9a5-48f0-801c-2e6aeed6070f/fd617927-f9a5-48f0-801c-2e6aeed6070f.vmdk to [datastore1] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1260.676562] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89eb94a2-fb71-4bf7-a2a2-40e91b06b5b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.699562] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.699941] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-798b64bf-6199-4151-b454-60d4857c2eb1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.719852] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1260.719852] env[62952]: value = "task-1367851" [ 1260.719852] env[62952]: _type = "Task" [ 1260.719852] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.728943] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367851, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.120369] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367850, 'name': ReconfigVM_Task, 'duration_secs': 0.513411} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.120796] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78/volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.126528] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c82b951-6113-4c80-9479-bc057a1da3e6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.142066] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1261.142066] env[62952]: value = "task-1367852" [ 1261.142066] env[62952]: _type = "Task" [ 1261.142066] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.150412] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367852, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.230387] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367851, 'name': ReconfigVM_Task, 'duration_secs': 0.370053} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.232058] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 27e78e78-3a7f-4ad4-b939-438cb2b90b55/27e78e78-3a7f-4ad4-b939-438cb2b90b55.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.232234] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encrypted': False, 'device_name': '/dev/sda', 'boot_index': 0, 'encryption_secret_uuid': None, 'encryption_options': None, 'guest_format': None, 'device_type': 'disk', 'encryption_format': None, 'size': 0, 'disk_bus': None, 'image_id': 'e04fcbd0-b3d8-461a-890d-eda04a2a1e77'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'attachment_id': 'd7f573f2-3630-4b2c-ac2d-b8b672bd4e43', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291003', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'name': 'volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '27e78e78-3a7f-4ad4-b939-438cb2b90b55', 'attached_at': '', 'detached_at': '', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'serial': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e'}, 'guest_format': None, 'device_type': None, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62952) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1261.232377] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Volume attach. Driver type: vmdk {{(pid=62952) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1261.232627] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291003', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'name': 'volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '27e78e78-3a7f-4ad4-b939-438cb2b90b55', 'attached_at': '', 'detached_at': '', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'serial': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1261.233651] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75547b48-e33e-4839-a834-5fe66491cbbd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.251524] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6176f193-9fd7-49ee-9321-3cd8994d923f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.276409] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e/volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1261.276738] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-877cd3ec-e92c-4c46-8695-c7c657091a5c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.295814] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1261.295814] env[62952]: value = "task-1367853" [ 1261.295814] env[62952]: _type = "Task" [ 1261.295814] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.303710] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367853, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.655181] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367852, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.806516] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367853, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.153687] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367852, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.307587] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367853, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.653508] env[62952]: DEBUG oslo_vmware.api [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367852, 'name': ReconfigVM_Task, 'duration_secs': 1.136933} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.653840] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291006', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'name': 'volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd7d45f3-31bb-43c3-9317-2142d29b6ef5', 'attached_at': '', 'detached_at': '', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'serial': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1262.815414] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367853, 'name': ReconfigVM_Task, 'duration_secs': 1.28923} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.815736] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e/volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e.vmdk or device None with type thin {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1262.821140] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72857e58-d84e-4812-a5e6-67de848d1baf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.842907] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1262.842907] env[62952]: value = "task-1367854" [ 1262.842907] env[62952]: _type = "Task" [ 1262.842907] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.855998] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367854, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.352477] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367854, 'name': ReconfigVM_Task, 'duration_secs': 0.173559} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.352804] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291003', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'name': 'volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '27e78e78-3a7f-4ad4-b939-438cb2b90b55', 'attached_at': '', 'detached_at': '', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'serial': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e'} {{(pid=62952) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1263.353455] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b18e4f1-30db-40ab-a6bd-29260e2c4a90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.360762] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1263.360762] env[62952]: value = "task-1367855" [ 1263.360762] env[62952]: _type = "Task" [ 1263.360762] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.369565] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367855, 'name': Rename_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.695833] env[62952]: DEBUG nova.objects.instance [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'flavor' on Instance uuid dd7d45f3-31bb-43c3-9317-2142d29b6ef5 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.871359] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367855, 'name': Rename_Task, 'duration_secs': 0.181584} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.871642] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1263.871924] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d23ada9-d3a7-4414-9c57-35748b41334a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.878527] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1263.878527] env[62952]: value = "task-1367856" [ 1263.878527] env[62952]: _type = "Task" [ 1263.878527] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.886026] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367856, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.951838] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.202592] env[62952]: DEBUG oslo_concurrency.lockutils [None req-74f2186b-a334-4328-a221-14e1736d48fc tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.769s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.203496] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.252s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.388709] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367856, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.706395] env[62952]: INFO nova.compute.manager [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Detaching volume 2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78 [ 1264.736431] env[62952]: INFO nova.virt.block_device [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Attempting to driver detach volume 2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78 from mountpoint /dev/sdb [ 1264.736681] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1264.736919] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291006', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'name': 'volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd7d45f3-31bb-43c3-9317-2142d29b6ef5', 'attached_at': '', 'detached_at': '', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'serial': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1264.737788] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b094a79-6db3-42b7-8b6e-9d5e1125eb90 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.759058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92242d58-6a37-4385-9805-0d4dd8ff84de {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.765895] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bd5b0f-8174-45c9-9602-b83695c18e68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.825137] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718c9069-78a4-4fe0-93e0-955e75433819 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.841766] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] The volume has not been displaced from its original location: [datastore1] volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78/volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1264.847063] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1264.847405] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e78f36cf-a9b8-4b83-bf4b-705f6cb617b9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.865267] env[62952]: DEBUG oslo_vmware.api [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1264.865267] env[62952]: value = "task-1367857" [ 1264.865267] env[62952]: _type = "Task" [ 1264.865267] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.873424] env[62952]: DEBUG oslo_vmware.api [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367857, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.887286] env[62952]: DEBUG oslo_vmware.api [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367856, 'name': PowerOnVM_Task, 'duration_secs': 0.521328} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.887541] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1264.999467] env[62952]: DEBUG nova.compute.manager [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1265.000587] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7c6abf-7e00-4f5a-9c87-f4d28cdcfb0a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.375675] env[62952]: DEBUG oslo_vmware.api [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367857, 'name': ReconfigVM_Task, 'duration_secs': 0.240643} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.376015] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1265.380752] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3779d9f4-efb4-40d4-9b6c-3bca710740bf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.396175] env[62952]: DEBUG oslo_vmware.api [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1265.396175] env[62952]: value = "task-1367858" [ 1265.396175] env[62952]: _type = "Task" [ 1265.396175] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.404403] env[62952]: DEBUG oslo_vmware.api [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367858, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.519100] env[62952]: DEBUG oslo_concurrency.lockutils [None req-63f13fb2-b091-4c8d-8845-4c3ce05f066f tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.844s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.714428] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.714628] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.905928] env[62952]: DEBUG oslo_vmware.api [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367858, 'name': ReconfigVM_Task, 'duration_secs': 0.150837} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.906274] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291006', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'name': 'volume-2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dd7d45f3-31bb-43c3-9317-2142d29b6ef5', 'attached_at': '', 'detached_at': '', 'volume_id': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78', 'serial': '2e8a9741-dc9c-4bed-9b7b-3bfb3b777e78'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1266.218164] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.218344] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.218534] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.218703] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1266.219711] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8442de0-706a-4a60-93fe-3f6f22608fdd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.228706] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdb1b9b-35de-4187-bb8a-158375495c2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.242598] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abb0898-cc9f-4e11-9825-7662b605781e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.249169] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c49b404-915c-4b3f-9374-47f754ffa18c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.279213] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180637MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1266.279793] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.279793] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.531591] env[62952]: DEBUG nova.objects.instance [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'flavor' on Instance uuid dd7d45f3-31bb-43c3-9317-2142d29b6ef5 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1267.269454] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5240b2b9-3279-c687-f1c6-c8d72982ddb4/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1267.270491] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e496abf-b902-46f4-823d-473967e70761 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.276872] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5240b2b9-3279-c687-f1c6-c8d72982ddb4/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1267.277084] env[62952]: ERROR oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5240b2b9-3279-c687-f1c6-c8d72982ddb4/disk-0.vmdk due to incomplete transfer. [ 1267.277314] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ad8de6f9-9a9c-4444-88a9-9c601be03811 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.284921] env[62952]: DEBUG oslo_vmware.rw_handles [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5240b2b9-3279-c687-f1c6-c8d72982ddb4/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1267.285140] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Uploaded image 2907e76b-65cb-4cbe-8115-6d1d708e65d3 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1267.287480] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1267.289956] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-876cd360-33fa-41a3-9d4c-939121f3bf1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.297844] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1267.297844] env[62952]: value = "task-1367859" [ 1267.297844] env[62952]: _type = "Task" [ 1267.297844] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.306328] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367859, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.309178] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance dd7d45f3-31bb-43c3-9317-2142d29b6ef5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.309362] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance c1c3d578-37e0-4823-9acf-612a09f512f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.309505] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 27e78e78-3a7f-4ad4-b939-438cb2b90b55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1267.309709] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1267.309858] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1267.360696] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bba3df-dac3-45fb-9f95-01f78f039149 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.367928] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b95765-d874-46a3-8313-711b32ad2f36 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.397115] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41519afd-6752-4de4-b988-81674fcc0861 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.404398] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27407c0a-ae60-492e-a781-a8a91824ac31 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.417298] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1267.542065] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ad6a95d7-78ef-4b62-bd1c-eaa5d7ffce5a tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.338s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.807872] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367859, 'name': Destroy_Task, 'duration_secs': 0.31015} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.808177] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Destroyed the VM [ 1267.808418] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1267.808673] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b7343062-90de-4b0e-9d05-855ad73b2211 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.815062] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1267.815062] env[62952]: value = "task-1367860" [ 1267.815062] env[62952]: _type = "Task" [ 1267.815062] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.822587] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367860, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.920706] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1268.198017] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.198313] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.198495] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.198679] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.198854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.200993] env[62952]: INFO nova.compute.manager [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Terminating instance [ 1268.202645] env[62952]: DEBUG nova.compute.manager [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1268.202862] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1268.203748] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acfc28e-855c-48bc-832e-6dd83e81356a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.212313] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1268.212543] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ecfb6cf3-b0f7-46fb-b87b-4ec930c06d7a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.219991] env[62952]: DEBUG oslo_vmware.api [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1268.219991] env[62952]: value = "task-1367861" [ 1268.219991] env[62952]: _type = "Task" [ 1268.219991] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.227951] env[62952]: DEBUG oslo_vmware.api [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.325891] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367860, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.426082] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1268.426370] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.147s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.730115] env[62952]: DEBUG oslo_vmware.api [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367861, 'name': PowerOffVM_Task, 'duration_secs': 0.201882} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.730532] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1268.730532] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1268.730701] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-496ef1cc-8a63-40d1-a816-b406069ac9b3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.799125] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1268.799372] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1268.799563] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleting the datastore file [datastore2] dd7d45f3-31bb-43c3-9317-2142d29b6ef5 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1268.799842] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a63e55a4-b768-415a-bd4d-b1491257e65f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.806767] env[62952]: DEBUG oslo_vmware.api [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for the task: (returnval){ [ 1268.806767] env[62952]: value = "task-1367863" [ 1268.806767] env[62952]: _type = "Task" [ 1268.806767] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.814660] env[62952]: DEBUG oslo_vmware.api [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367863, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.824752] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367860, 'name': RemoveSnapshot_Task, 'duration_secs': 0.54106} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.824843] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1268.825197] env[62952]: DEBUG nova.compute.manager [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1268.826140] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ecc397-63d2-44cb-b2a4-ec386cb867a7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.316856] env[62952]: DEBUG oslo_vmware.api [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Task: {'id': task-1367863, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130892} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.318067] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1269.318067] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1269.318067] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1269.318067] env[62952]: INFO nova.compute.manager [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1269.318317] env[62952]: DEBUG oslo.service.loopingcall [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1269.318408] env[62952]: DEBUG nova.compute.manager [-] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1269.318503] env[62952]: DEBUG nova.network.neutron [-] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1269.339159] env[62952]: INFO nova.compute.manager [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Shelve offloading [ 1269.340688] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1269.340923] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4509c88b-89e3-4c81-bda3-d2d13e02eeb4 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.348341] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1269.348341] env[62952]: value = "task-1367864" [ 1269.348341] env[62952]: _type = "Task" [ 1269.348341] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.357314] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.426664] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.426914] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.427106] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.427265] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.427406] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1269.714468] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.714647] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1269.798978] env[62952]: DEBUG nova.compute.manager [req-a176d9da-f36b-45ef-bee3-251d876daf0e req-ede32f76-baca-4a6e-ad85-5833499565f1 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Received event network-vif-deleted-199c0f83-9683-4568-806b-f01b051f2f0c {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1269.799289] env[62952]: INFO nova.compute.manager [req-a176d9da-f36b-45ef-bee3-251d876daf0e req-ede32f76-baca-4a6e-ad85-5833499565f1 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Neutron deleted interface 199c0f83-9683-4568-806b-f01b051f2f0c; detaching it from the instance and deleting it from the info cache [ 1269.799289] env[62952]: DEBUG nova.network.neutron [req-a176d9da-f36b-45ef-bee3-251d876daf0e req-ede32f76-baca-4a6e-ad85-5833499565f1 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1269.860549] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1269.860711] env[62952]: DEBUG nova.compute.manager [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1269.861514] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82225031-290b-4523-8f2d-b0da4fb51b96 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.867686] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.867855] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.868039] env[62952]: DEBUG nova.network.neutron [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1270.217295] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1270.217556] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1270.276247] env[62952]: DEBUG nova.network.neutron [-] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.301234] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f0656dc-f3b7-4b22-8a6b-393983fe9c41 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.311977] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3603595d-f255-46bd-ab0b-54ca0025b55f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.337268] env[62952]: DEBUG nova.compute.manager [req-a176d9da-f36b-45ef-bee3-251d876daf0e req-ede32f76-baca-4a6e-ad85-5833499565f1 service nova] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Detach interface failed, port_id=199c0f83-9683-4568-806b-f01b051f2f0c, reason: Instance dd7d45f3-31bb-43c3-9317-2142d29b6ef5 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1270.575365] env[62952]: DEBUG nova.network.neutron [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.714210] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1270.778872] env[62952]: INFO nova.compute.manager [-] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Took 1.46 seconds to deallocate network for instance. [ 1271.078558] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1271.285466] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.285820] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.286067] env[62952]: DEBUG nova.objects.instance [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lazy-loading 'resources' on Instance uuid dd7d45f3-31bb-43c3-9317-2142d29b6ef5 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1271.384267] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1271.385221] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fb1fc7-737b-482f-b424-d6a80ec14b6e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.394224] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1271.394490] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99be206d-e57a-4410-ba3e-e915d88bf725 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.459551] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1271.459903] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1271.460163] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleting the datastore file [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1271.460443] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b487fef-18ef-493a-aa7a-5e57fdf9f08e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.468230] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1271.468230] env[62952]: value = "task-1367866" [ 1271.468230] env[62952]: _type = "Task" [ 1271.468230] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.478850] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.710567] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.823906] env[62952]: DEBUG nova.compute.manager [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-vif-unplugged-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1271.824156] env[62952]: DEBUG oslo_concurrency.lockutils [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.824372] env[62952]: DEBUG oslo_concurrency.lockutils [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.824565] env[62952]: DEBUG oslo_concurrency.lockutils [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.824747] env[62952]: DEBUG nova.compute.manager [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] No waiting events found dispatching network-vif-unplugged-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1271.824940] env[62952]: WARNING nova.compute.manager [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received unexpected event network-vif-unplugged-208e022a-eaea-4c27-ac52-5773ba975aa3 for instance with vm_state shelved and task_state shelving_offloading. [ 1271.825131] env[62952]: DEBUG nova.compute.manager [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1271.825363] env[62952]: DEBUG nova.compute.manager [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing instance network info cache due to event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1271.826015] env[62952]: DEBUG oslo_concurrency.lockutils [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1271.826015] env[62952]: DEBUG oslo_concurrency.lockutils [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1271.826015] env[62952]: DEBUG nova.network.neutron [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1271.851899] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b5b5f0-54dc-4d5e-bcc6-a9d38d673437 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.860249] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff5d9a1-f948-415c-9b6e-a145beff9281 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.889972] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40590bf-8a80-4147-a5e6-43f638e2dbbd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.897465] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3106b9-2db6-4b2c-9e63-3926ee655f03 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.910411] env[62952]: DEBUG nova.compute.provider_tree [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.978559] env[62952]: DEBUG oslo_vmware.api [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124123} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.978920] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1271.979144] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1271.979328] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1271.999033] env[62952]: INFO nova.scheduler.client.report [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted allocations for instance c1c3d578-37e0-4823-9acf-612a09f512f7 [ 1272.413460] env[62952]: DEBUG nova.scheduler.client.report [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1272.502230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.537772] env[62952]: DEBUG nova.network.neutron [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updated VIF entry in instance network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1272.538198] env[62952]: DEBUG nova.network.neutron [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": null, "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap208e022a-ea", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1272.918284] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.632s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.921027] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.419s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.921027] env[62952]: DEBUG nova.objects.instance [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'resources' on Instance uuid c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1272.937679] env[62952]: INFO nova.scheduler.client.report [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Deleted allocations for instance dd7d45f3-31bb-43c3-9317-2142d29b6ef5 [ 1273.040888] env[62952]: DEBUG oslo_concurrency.lockutils [req-aaedaf75-3c9f-4b68-9471-878664b1bd13 req-d97944c4-6078-4a5f-b31b-5bbc484bcf4d service nova] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1273.427204] env[62952]: DEBUG nova.objects.instance [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'numa_topology' on Instance uuid c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1273.443927] env[62952]: DEBUG oslo_concurrency.lockutils [None req-460bbc8f-2eb5-421f-92bb-b27d0695f2d4 tempest-AttachVolumeNegativeTest-1355480005 tempest-AttachVolumeNegativeTest-1355480005-project-member] Lock "dd7d45f3-31bb-43c3-9317-2142d29b6ef5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.246s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.929464] env[62952]: DEBUG nova.objects.base [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1273.966685] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e0f9f5-9764-4a2b-8fc5-c8d83eeae6c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.974928] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aba89f7-6b94-4231-9a20-b4819e9c92b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.004483] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5673a01c-6686-40da-8f72-785d6f44bc83 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.011824] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02c2e3d-6871-468f-af08-1a360d2866dd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.024647] env[62952]: DEBUG nova.compute.provider_tree [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1274.342308] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1274.527977] env[62952]: DEBUG nova.scheduler.client.report [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1275.032635] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.541113] env[62952]: DEBUG oslo_concurrency.lockutils [None req-d932ef30-3f50-4852-8fb6-1cf388f806ef tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.317s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.541974] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.200s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.542189] env[62952]: INFO nova.compute.manager [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Unshelving [ 1276.566490] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.566757] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.566969] env[62952]: DEBUG nova.objects.instance [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'pci_requests' on Instance uuid c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.071657] env[62952]: DEBUG nova.objects.instance [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'numa_topology' on Instance uuid c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.574020] env[62952]: INFO nova.compute.claims [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1278.623365] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531af08a-0bf0-4372-b2de-1c733b1814cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.631157] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66890a8-cfa7-43ab-bae3-5353df0f5427 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.660765] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8089e71b-0a69-4ba0-b5bd-d4a812b9680f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.667691] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b184c1bd-a6a2-42b4-92cb-a5ec10a59f2a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.681655] env[62952]: DEBUG nova.compute.provider_tree [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1279.185372] env[62952]: DEBUG nova.scheduler.client.report [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1279.690705] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.124s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.717782] env[62952]: INFO nova.network.neutron [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating port 208e022a-eaea-4c27-ac52-5773ba975aa3 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1281.230920] env[62952]: DEBUG nova.compute.manager [req-2eadc225-fb6a-4a59-97d0-1b784222be55 req-155c0deb-d2aa-4ea0-8b98-09b0b837807f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-vif-plugged-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1281.232367] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eadc225-fb6a-4a59-97d0-1b784222be55 req-155c0deb-d2aa-4ea0-8b98-09b0b837807f service nova] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.232676] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eadc225-fb6a-4a59-97d0-1b784222be55 req-155c0deb-d2aa-4ea0-8b98-09b0b837807f service nova] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.232909] env[62952]: DEBUG oslo_concurrency.lockutils [req-2eadc225-fb6a-4a59-97d0-1b784222be55 req-155c0deb-d2aa-4ea0-8b98-09b0b837807f service nova] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.233138] env[62952]: DEBUG nova.compute.manager [req-2eadc225-fb6a-4a59-97d0-1b784222be55 req-155c0deb-d2aa-4ea0-8b98-09b0b837807f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] No waiting events found dispatching network-vif-plugged-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1281.233445] env[62952]: WARNING nova.compute.manager [req-2eadc225-fb6a-4a59-97d0-1b784222be55 req-155c0deb-d2aa-4ea0-8b98-09b0b837807f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received unexpected event network-vif-plugged-208e022a-eaea-4c27-ac52-5773ba975aa3 for instance with vm_state shelved_offloaded and task_state spawning. [ 1281.398847] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.399059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.399249] env[62952]: DEBUG nova.network.neutron [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1282.250878] env[62952]: DEBUG nova.network.neutron [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1282.758048] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.790915] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='10c55b5eae367888a79c863926300745',container_format='bare',created_at=2024-10-26T01:45:00Z,direct_url=,disk_format='vmdk',id=2907e76b-65cb-4cbe-8115-6d1d708e65d3,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-297069895-shelved',owner='0c451895447b473eb18675f6bc843081',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-26T01:45:14Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1282.790915] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1282.790915] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1282.791355] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1282.791659] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1282.791958] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1282.792306] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1282.792591] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1282.792901] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1282.793210] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1282.793532] env[62952]: DEBUG nova.virt.hardware [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1282.794532] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10247b5e-d6f5-4adc-bf69-6607f3cc1e17 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.803356] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b69929-6175-4711-8045-8165c6596862 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.818754] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:fd:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fb6e5b3-7be7-4018-8d9e-2133d926178c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '208e022a-eaea-4c27-ac52-5773ba975aa3', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1282.826113] env[62952]: DEBUG oslo.service.loopingcall [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1282.826302] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1282.826539] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77538c56-af01-493e-859b-e796de14ab1d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.847391] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1282.847391] env[62952]: value = "task-1367870" [ 1282.847391] env[62952]: _type = "Task" [ 1282.847391] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.551140] env[62952]: DEBUG nova.compute.manager [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1283.551496] env[62952]: DEBUG nova.compute.manager [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing instance network info cache due to event network-changed-208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1283.551496] env[62952]: DEBUG oslo_concurrency.lockutils [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] Acquiring lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.551638] env[62952]: DEBUG oslo_concurrency.lockutils [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] Acquired lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.551804] env[62952]: DEBUG nova.network.neutron [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Refreshing network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1283.558197] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367870, 'name': CreateVM_Task, 'duration_secs': 0.307913} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.558338] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1283.558909] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.559082] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.559489] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1283.560158] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fedf056a-3ff5-43b5-9978-0f2be999f0d7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.564702] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1283.564702] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5236ee83-dbfd-ff8d-28a1-8278c268edfe" [ 1283.564702] env[62952]: _type = "Task" [ 1283.564702] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.572595] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5236ee83-dbfd-ff8d-28a1-8278c268edfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.074458] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.074705] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Processing image 2907e76b-65cb-4cbe-8115-6d1d708e65d3 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1284.074945] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.075127] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.075312] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1284.075579] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0899fe73-0d07-4e52-a03c-22713a573de8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.091918] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1284.092110] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1284.092809] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca04009-1293-4bcc-a646-0f0ebb734d68 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.098023] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1284.098023] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52de79ed-bb43-713b-d735-7346cbfeb905" [ 1284.098023] env[62952]: _type = "Task" [ 1284.098023] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.104961] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52de79ed-bb43-713b-d735-7346cbfeb905, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.264994] env[62952]: DEBUG nova.network.neutron [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updated VIF entry in instance network info cache for port 208e022a-eaea-4c27-ac52-5773ba975aa3. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1284.265462] env[62952]: DEBUG nova.network.neutron [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [{"id": "208e022a-eaea-4c27-ac52-5773ba975aa3", "address": "fa:16:3e:e0:fd:c8", "network": {"id": "82888eba-739e-4bb6-b1d0-345fbcc0742f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-497652438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c451895447b473eb18675f6bc843081", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fb6e5b3-7be7-4018-8d9e-2133d926178c", "external-id": "nsx-vlan-transportzone-188", "segmentation_id": 188, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap208e022a-ea", "ovs_interfaceid": "208e022a-eaea-4c27-ac52-5773ba975aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.608573] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1284.608986] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Fetch image to [datastore2] OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39/OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1284.609046] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Downloading stream optimized image 2907e76b-65cb-4cbe-8115-6d1d708e65d3 to [datastore2] OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39/OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39.vmdk on the data store datastore2 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1284.609229] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Downloading image file data 2907e76b-65cb-4cbe-8115-6d1d708e65d3 to the ESX as VM named 'OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1284.679640] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1284.679640] env[62952]: value = "resgroup-9" [ 1284.679640] env[62952]: _type = "ResourcePool" [ 1284.679640] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1284.679986] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b16de3f8-2f73-4bb0-b549-dd599ff80201 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.702142] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease: (returnval){ [ 1284.702142] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526d14d6-3ec2-ecf8-0f69-10e7462d45d2" [ 1284.702142] env[62952]: _type = "HttpNfcLease" [ 1284.702142] env[62952]: } obtained for vApp import into resource pool (val){ [ 1284.702142] env[62952]: value = "resgroup-9" [ 1284.702142] env[62952]: _type = "ResourcePool" [ 1284.702142] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1284.702422] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the lease: (returnval){ [ 1284.702422] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526d14d6-3ec2-ecf8-0f69-10e7462d45d2" [ 1284.702422] env[62952]: _type = "HttpNfcLease" [ 1284.702422] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1284.708242] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1284.708242] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526d14d6-3ec2-ecf8-0f69-10e7462d45d2" [ 1284.708242] env[62952]: _type = "HttpNfcLease" [ 1284.708242] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1284.768200] env[62952]: DEBUG oslo_concurrency.lockutils [req-0d708d99-227b-4c32-b03f-99586eaac388 req-df298348-ee97-42ce-899b-fef977c6a8f2 service nova] Releasing lock "refresh_cache-c1c3d578-37e0-4823-9acf-612a09f512f7" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.210597] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1285.210597] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526d14d6-3ec2-ecf8-0f69-10e7462d45d2" [ 1285.210597] env[62952]: _type = "HttpNfcLease" [ 1285.210597] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1285.211053] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1285.211053] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]526d14d6-3ec2-ecf8-0f69-10e7462d45d2" [ 1285.211053] env[62952]: _type = "HttpNfcLease" [ 1285.211053] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1285.211600] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29de8e30-f2c1-408e-bc02-f24057bd0680 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.218765] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52495e1c-93bd-ed78-834d-5ecaa47e2ce9/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1285.218948] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52495e1c-93bd-ed78-834d-5ecaa47e2ce9/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1285.313626] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ca3981ee-b674-45d5-858b-3fb9ca083b2e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.370430] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1286.370928] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52495e1c-93bd-ed78-834d-5ecaa47e2ce9/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1286.371785] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2af0f1-65da-4749-800f-eba8ec814167 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.378821] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52495e1c-93bd-ed78-834d-5ecaa47e2ce9/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1286.378987] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52495e1c-93bd-ed78-834d-5ecaa47e2ce9/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1286.379224] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e2d8084e-958d-42fc-bbfd-13aff9f750a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.575169] env[62952]: DEBUG oslo_vmware.rw_handles [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52495e1c-93bd-ed78-834d-5ecaa47e2ce9/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1286.575413] env[62952]: INFO nova.virt.vmwareapi.images [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Downloaded image file data 2907e76b-65cb-4cbe-8115-6d1d708e65d3 [ 1286.576277] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41124ef0-d574-4c0c-962e-ad52b40a9773 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.592559] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-181faf79-4e0d-47d9-b724-963bd7bf87be {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.614067] env[62952]: INFO nova.virt.vmwareapi.images [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] The imported VM was unregistered [ 1286.616211] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1286.616475] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1286.616732] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17dfc5db-91a7-4b6b-9c00-495e42d6fd86 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.634332] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Created directory with path [datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1286.634454] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39/OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39.vmdk to [datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1286.634683] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0e7bd664-56fe-44b8-970e-d572cba92bee {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.640849] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1286.640849] env[62952]: value = "task-1367873" [ 1286.640849] env[62952]: _type = "Task" [ 1286.640849] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.647975] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367873, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.150561] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367873, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.650809] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367873, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.152228] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367873, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.653576] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367873, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.153811] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367873, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.197074} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.153811] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39/OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39.vmdk to [datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk. [ 1289.154072] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Cleaning up location [datastore2] OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1289.154072] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_7b32da39-23d4-4d0c-a39b-79b4c14dda39 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1289.154271] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc61b8fc-7762-4cf6-8de3-6a64fca4b84c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.159697] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1289.159697] env[62952]: value = "task-1367874" [ 1289.159697] env[62952]: _type = "Task" [ 1289.159697] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.166567] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367874, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.670249] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367874, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03603} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.670701] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1289.670701] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.670943] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk to [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1289.671208] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a02d72ce-76ea-4749-8a84-20b58f00e234 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.677762] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1289.677762] env[62952]: value = "task-1367875" [ 1289.677762] env[62952]: _type = "Task" [ 1289.677762] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.685965] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.187436] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.689440] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.189567] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.690561] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.190742] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.691838] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367875, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.523216} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.692237] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2907e76b-65cb-4cbe-8115-6d1d708e65d3/2907e76b-65cb-4cbe-8115-6d1d708e65d3.vmdk to [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1292.692927] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa860fe0-6f2f-4d73-9f94-e634121d9caa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.714634] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1292.714863] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a38e0242-a459-4294-ae7c-7723eec82454 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.732668] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1292.732668] env[62952]: value = "task-1367876" [ 1292.732668] env[62952]: _type = "Task" [ 1292.732668] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.739728] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367876, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.242361] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367876, 'name': ReconfigVM_Task, 'duration_secs': 0.273845} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.242678] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Reconfigured VM instance instance-0000006e to attach disk [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7/c1c3d578-37e0-4823-9acf-612a09f512f7.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1293.243356] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cefa79a5-33e9-44b1-b6d7-ef1b492968c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.250053] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1293.250053] env[62952]: value = "task-1367877" [ 1293.250053] env[62952]: _type = "Task" [ 1293.250053] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.257672] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367877, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.759312] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367877, 'name': Rename_Task, 'duration_secs': 0.146352} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.759688] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1293.759859] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-357931cf-4ba4-4d77-a9cf-a7030717f987 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.765978] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1293.765978] env[62952]: value = "task-1367878" [ 1293.765978] env[62952]: _type = "Task" [ 1293.765978] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.773195] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367878, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.276252] env[62952]: DEBUG oslo_vmware.api [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367878, 'name': PowerOnVM_Task, 'duration_secs': 0.426795} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.276522] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1294.378027] env[62952]: DEBUG nova.compute.manager [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1294.378512] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87bfd09a-bc59-4a30-afcb-6e38a0b97771 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.895596] env[62952]: DEBUG oslo_concurrency.lockutils [None req-ea4f1648-0643-4f29-b8f6-f72b0d9e24af tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.354s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.418045] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.418045] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.418296] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.418452] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.418624] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.420722] env[62952]: INFO nova.compute.manager [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Terminating instance [ 1295.422384] env[62952]: DEBUG nova.compute.manager [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1295.422580] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1295.423420] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5259fa2a-f6d6-431c-a309-a6283459fa07 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.430576] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1295.430787] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4bfcacff-5173-4c71-86a3-1b88baabd641 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.437011] env[62952]: DEBUG oslo_vmware.api [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1295.437011] env[62952]: value = "task-1367879" [ 1295.437011] env[62952]: _type = "Task" [ 1295.437011] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.444309] env[62952]: DEBUG oslo_vmware.api [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.947475] env[62952]: DEBUG oslo_vmware.api [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367879, 'name': PowerOffVM_Task, 'duration_secs': 0.177167} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.947918] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1295.947918] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1295.948162] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fbb55487-bec3-4275-9a32-37cfe7b99712 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.008528] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1296.008758] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1296.008943] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleting the datastore file [datastore2] c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1296.009223] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3142772-c532-40a2-949c-35d2ff5c458e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.015112] env[62952]: DEBUG oslo_vmware.api [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for the task: (returnval){ [ 1296.015112] env[62952]: value = "task-1367881" [ 1296.015112] env[62952]: _type = "Task" [ 1296.015112] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.022242] env[62952]: DEBUG oslo_vmware.api [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.524475] env[62952]: DEBUG oslo_vmware.api [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Task: {'id': task-1367881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148162} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.524723] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1296.524906] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1296.525096] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1296.525296] env[62952]: INFO nova.compute.manager [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1296.525540] env[62952]: DEBUG oslo.service.loopingcall [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1296.525728] env[62952]: DEBUG nova.compute.manager [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1296.525820] env[62952]: DEBUG nova.network.neutron [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1296.951131] env[62952]: DEBUG nova.compute.manager [req-d7f81416-358c-4504-8639-bf4a16a83239 req-fadb8391-f1ba-4b71-890a-283d2decf77f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Received event network-vif-deleted-208e022a-eaea-4c27-ac52-5773ba975aa3 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1296.951131] env[62952]: INFO nova.compute.manager [req-d7f81416-358c-4504-8639-bf4a16a83239 req-fadb8391-f1ba-4b71-890a-283d2decf77f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Neutron deleted interface 208e022a-eaea-4c27-ac52-5773ba975aa3; detaching it from the instance and deleting it from the info cache [ 1296.951131] env[62952]: DEBUG nova.network.neutron [req-d7f81416-358c-4504-8639-bf4a16a83239 req-fadb8391-f1ba-4b71-890a-283d2decf77f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1297.422600] env[62952]: DEBUG nova.network.neutron [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1297.452906] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd4aadaf-7567-4fcf-9f5b-84c38e277b9e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.462983] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297dd2c1-60b8-4244-ba08-06da51facb4f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.486121] env[62952]: DEBUG nova.compute.manager [req-d7f81416-358c-4504-8639-bf4a16a83239 req-fadb8391-f1ba-4b71-890a-283d2decf77f service nova] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Detach interface failed, port_id=208e022a-eaea-4c27-ac52-5773ba975aa3, reason: Instance c1c3d578-37e0-4823-9acf-612a09f512f7 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1297.924927] env[62952]: INFO nova.compute.manager [-] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Took 1.40 seconds to deallocate network for instance. [ 1298.432154] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.432531] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.432630] env[62952]: DEBUG nova.objects.instance [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lazy-loading 'resources' on Instance uuid c1c3d578-37e0-4823-9acf-612a09f512f7 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1298.982578] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb890cd1-e611-44c5-82ef-eb93bdba4fed {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.989472] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd60f45-1f7d-45c1-bb6e-f06cbbd9b583 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.019913] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ce7756-b4ea-46d4-953f-e7374f625865 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.027058] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a39626-c7cd-4217-a3a4-6bb86b5b97cd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.040039] env[62952]: DEBUG nova.compute.provider_tree [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1299.546074] env[62952]: DEBUG nova.scheduler.client.report [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1300.050056] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1300.071741] env[62952]: INFO nova.scheduler.client.report [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Deleted allocations for instance c1c3d578-37e0-4823-9acf-612a09f512f7 [ 1300.579521] env[62952]: DEBUG oslo_concurrency.lockutils [None req-a73438a7-0a30-45a6-9f67-fff7634541e0 tempest-ServerActionsTestOtherB-1236125099 tempest-ServerActionsTestOtherB-1236125099-project-member] Lock "c1c3d578-37e0-4823-9acf-612a09f512f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.161s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1302.182230] env[62952]: DEBUG oslo_concurrency.lockutils [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.182580] env[62952]: DEBUG oslo_concurrency.lockutils [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1302.686193] env[62952]: INFO nova.compute.manager [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Detaching volume ca518c35-2fc1-4cac-b6d5-c2c01d5b057e [ 1302.719555] env[62952]: INFO nova.virt.block_device [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Attempting to driver detach volume ca518c35-2fc1-4cac-b6d5-c2c01d5b057e from mountpoint /dev/sdb [ 1302.719781] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Volume detach. Driver type: vmdk {{(pid=62952) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1302.719974] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291003', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'name': 'volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '27e78e78-3a7f-4ad4-b939-438cb2b90b55', 'attached_at': '', 'detached_at': '', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'serial': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1302.720874] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5e31f3-e0c0-42ce-97aa-8b7d67efd854 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.742247] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ec1bd2-84bd-4cbd-a9e8-166c004d8c60 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.750883] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f22dcfa-8869-4d78-8b57-97d0208e5a1c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.771253] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07e4d98-ff6f-46e7-b59e-efaaf7ca3b6a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.786322] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] The volume has not been displaced from its original location: [datastore1] volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e/volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e.vmdk. No consolidation needed. {{(pid=62952) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1302.791495] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1302.791798] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7101c515-00f5-4f00-b608-826aacb7b29f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.814127] env[62952]: DEBUG oslo_vmware.api [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1302.814127] env[62952]: value = "task-1367883" [ 1302.814127] env[62952]: _type = "Task" [ 1302.814127] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.823087] env[62952]: DEBUG oslo_vmware.api [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.324474] env[62952]: DEBUG oslo_vmware.api [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367883, 'name': ReconfigVM_Task, 'duration_secs': 0.215521} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.324874] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=62952) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1303.329533] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c3a3421-9f0d-4de4-83bc-fef2e5e4d640 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.344149] env[62952]: DEBUG oslo_vmware.api [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1303.344149] env[62952]: value = "task-1367884" [ 1303.344149] env[62952]: _type = "Task" [ 1303.344149] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.351779] env[62952]: DEBUG oslo_vmware.api [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367884, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.854898] env[62952]: DEBUG oslo_vmware.api [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367884, 'name': ReconfigVM_Task, 'duration_secs': 0.144175} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.855326] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-291003', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'name': 'volume-ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '27e78e78-3a7f-4ad4-b939-438cb2b90b55', 'attached_at': '', 'detached_at': '', 'volume_id': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e', 'serial': 'ca518c35-2fc1-4cac-b6d5-c2c01d5b057e'} {{(pid=62952) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1304.404418] env[62952]: DEBUG nova.objects.instance [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'flavor' on Instance uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1305.414656] env[62952]: DEBUG oslo_concurrency.lockutils [None req-83872854-0743-4959-a104-1fa320e97450 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.232s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.451952] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.452399] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1306.452447] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.452675] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1306.452913] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.455260] env[62952]: INFO nova.compute.manager [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Terminating instance [ 1306.457118] env[62952]: DEBUG nova.compute.manager [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1306.457357] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1306.458246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06504f9f-15a2-41b4-ab17-301055a7c0c0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.467118] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1306.467379] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a75225d8-b000-4cbc-83ae-4dbbe96c71fa {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.472682] env[62952]: DEBUG oslo_vmware.api [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1306.472682] env[62952]: value = "task-1367885" [ 1306.472682] env[62952]: _type = "Task" [ 1306.472682] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.480210] env[62952]: DEBUG oslo_vmware.api [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.982497] env[62952]: DEBUG oslo_vmware.api [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367885, 'name': PowerOffVM_Task, 'duration_secs': 0.184367} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.982768] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1306.982939] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1306.983227] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e54dbfe3-7cf0-407f-a11e-36dfe879c0c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.045759] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1307.045980] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deleting contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1307.046178] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleting the datastore file [datastore1] 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1307.046432] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2114ce4-b7d7-4e11-ba44-0287505af810 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.052650] env[62952]: DEBUG oslo_vmware.api [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1307.052650] env[62952]: value = "task-1367887" [ 1307.052650] env[62952]: _type = "Task" [ 1307.052650] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.060705] env[62952]: DEBUG oslo_vmware.api [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.562368] env[62952]: DEBUG oslo_vmware.api [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143192} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.562849] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1307.562849] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deleted contents of the VM from datastore datastore1 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1307.562955] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1307.563190] env[62952]: INFO nova.compute.manager [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1307.563444] env[62952]: DEBUG oslo.service.loopingcall [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1307.563680] env[62952]: DEBUG nova.compute.manager [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1307.563804] env[62952]: DEBUG nova.network.neutron [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1307.978090] env[62952]: DEBUG nova.compute.manager [req-de788d8e-5bbf-4628-8638-c0a6428bd4f7 req-2ffb3fd7-96c9-4772-a165-7aecb62d0ed6 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Received event network-vif-deleted-551a3758-07da-4c7a-acda-583eb14f995f {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1307.978325] env[62952]: INFO nova.compute.manager [req-de788d8e-5bbf-4628-8638-c0a6428bd4f7 req-2ffb3fd7-96c9-4772-a165-7aecb62d0ed6 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Neutron deleted interface 551a3758-07da-4c7a-acda-583eb14f995f; detaching it from the instance and deleting it from the info cache [ 1307.978504] env[62952]: DEBUG nova.network.neutron [req-de788d8e-5bbf-4628-8638-c0a6428bd4f7 req-2ffb3fd7-96c9-4772-a165-7aecb62d0ed6 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1308.451014] env[62952]: DEBUG nova.network.neutron [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1308.481949] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a468546c-5e21-4b6b-a2f2-60092da51673 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.491877] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2c4ae0-a6c3-4944-acaa-c772ab71c561 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.519878] env[62952]: DEBUG nova.compute.manager [req-de788d8e-5bbf-4628-8638-c0a6428bd4f7 req-2ffb3fd7-96c9-4772-a165-7aecb62d0ed6 service nova] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Detach interface failed, port_id=551a3758-07da-4c7a-acda-583eb14f995f, reason: Instance 27e78e78-3a7f-4ad4-b939-438cb2b90b55 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1308.953314] env[62952]: INFO nova.compute.manager [-] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Took 1.39 seconds to deallocate network for instance. [ 1309.461088] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.461382] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1309.461614] env[62952]: DEBUG nova.objects.instance [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'resources' on Instance uuid 27e78e78-3a7f-4ad4-b939-438cb2b90b55 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.997257] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39916716-6431-424a-9a31-ca32b29b6098 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.004675] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328728d1-be39-42c3-bc4e-c689270bd61f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.034123] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07eed067-96ff-4307-bd0e-137a186c116a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.040555] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a095a4-353e-4ab3-911d-78b3ccf3dde0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.053109] env[62952]: DEBUG nova.compute.provider_tree [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.556770] env[62952]: DEBUG nova.scheduler.client.report [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1311.061411] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.079488] env[62952]: INFO nova.scheduler.client.report [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted allocations for instance 27e78e78-3a7f-4ad4-b939-438cb2b90b55 [ 1311.587888] env[62952]: DEBUG oslo_concurrency.lockutils [None req-e2aedd98-5dda-43b8-8426-26ac7f58c8aa tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "27e78e78-3a7f-4ad4-b939-438cb2b90b55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.136s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.246600] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.246894] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.748772] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Starting instance... {{(pid=62952) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1314.272690] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.272934] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.274456] env[62952]: INFO nova.compute.claims [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1315.307611] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80492a6e-523a-425b-88b5-cf062bbc68b8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.314531] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2097036c-4fff-445a-a0d6-b94db0071933 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.342995] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1b8da0-838b-4e57-8e1a-39f5a8bb4c63 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.349562] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa68acd-360f-4e2e-9184-04f4c1412cb7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.362103] env[62952]: DEBUG nova.compute.provider_tree [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1315.864846] env[62952]: DEBUG nova.scheduler.client.report [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1316.369951] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.097s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1316.370482] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Start building networks asynchronously for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1316.875667] env[62952]: DEBUG nova.compute.utils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Using /dev/sd instead of None {{(pid=62952) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1316.877144] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Allocating IP information in the background. {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1316.877330] env[62952]: DEBUG nova.network.neutron [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] allocate_for_instance() {{(pid=62952) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1316.928781] env[62952]: DEBUG nova.policy [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77db11499b634ddf847014f4f4c169f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2c2a97031df4b75b835a96b41793c74', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62952) authorize /opt/stack/nova/nova/policy.py:201}} [ 1317.183842] env[62952]: DEBUG nova.network.neutron [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Successfully created port: 9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1317.380383] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Start building block device mappings for instance. {{(pid=62952) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1318.391671] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Start spawning the instance on the hypervisor. {{(pid=62952) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1318.416157] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-26T01:31:58Z,direct_url=,disk_format='vmdk',id=e04fcbd0-b3d8-461a-890d-eda04a2a1e77,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='30d8f6806bb74841862ef4ef6a4490a1',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-26T01:31:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1318.416444] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1318.416607] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1318.416791] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1318.416939] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1318.417103] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1318.417314] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1318.417473] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1318.417639] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1318.417800] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1318.417974] env[62952]: DEBUG nova.virt.hardware [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.418893] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d510b875-2d03-42bb-8958-99a573ee6b30 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.426562] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a341362-595f-4060-8136-6acc54fa6143 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.543744] env[62952]: DEBUG nova.compute.manager [req-f33961b1-c6ad-402a-b9b1-c7b02b5f4581 req-f0912b42-c86b-4c1a-aded-d669ada63579 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-vif-plugged-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1318.543976] env[62952]: DEBUG oslo_concurrency.lockutils [req-f33961b1-c6ad-402a-b9b1-c7b02b5f4581 req-f0912b42-c86b-4c1a-aded-d669ada63579 service nova] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1318.544210] env[62952]: DEBUG oslo_concurrency.lockutils [req-f33961b1-c6ad-402a-b9b1-c7b02b5f4581 req-f0912b42-c86b-4c1a-aded-d669ada63579 service nova] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1318.544387] env[62952]: DEBUG oslo_concurrency.lockutils [req-f33961b1-c6ad-402a-b9b1-c7b02b5f4581 req-f0912b42-c86b-4c1a-aded-d669ada63579 service nova] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1318.544557] env[62952]: DEBUG nova.compute.manager [req-f33961b1-c6ad-402a-b9b1-c7b02b5f4581 req-f0912b42-c86b-4c1a-aded-d669ada63579 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] No waiting events found dispatching network-vif-plugged-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1318.544729] env[62952]: WARNING nova.compute.manager [req-f33961b1-c6ad-402a-b9b1-c7b02b5f4581 req-f0912b42-c86b-4c1a-aded-d669ada63579 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received unexpected event network-vif-plugged-9b7beec0-076c-404c-864f-f225c46c9ac2 for instance with vm_state building and task_state spawning. [ 1318.623696] env[62952]: DEBUG nova.network.neutron [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Successfully updated port: 9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1319.127055] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.127055] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.127205] env[62952]: DEBUG nova.network.neutron [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1319.657199] env[62952]: DEBUG nova.network.neutron [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance cache missing network info. {{(pid=62952) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1319.774170] env[62952]: DEBUG nova.network.neutron [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1320.276708] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.276998] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance network_info: |[{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62952) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1320.277490] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:0a:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd22cb4ec-277f-41ee-8aba-b3d54442b93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b7beec0-076c-404c-864f-f225c46c9ac2', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1320.284945] env[62952]: DEBUG oslo.service.loopingcall [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1320.285195] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1320.285428] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cf27128-3fa7-4d2e-9867-5fcae89efe08 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.305659] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1320.305659] env[62952]: value = "task-1367888" [ 1320.305659] env[62952]: _type = "Task" [ 1320.305659] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.312836] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367888, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.574804] env[62952]: DEBUG nova.compute.manager [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1320.575068] env[62952]: DEBUG nova.compute.manager [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing instance network info cache due to event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1320.575311] env[62952]: DEBUG oslo_concurrency.lockutils [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1320.575461] env[62952]: DEBUG oslo_concurrency.lockutils [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1320.575628] env[62952]: DEBUG nova.network.neutron [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1320.814824] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367888, 'name': CreateVM_Task, 'duration_secs': 0.292934} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.815206] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1320.815727] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1320.815912] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1320.816267] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1320.816517] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b611ad31-58ef-48a4-bb7c-48885acf3197 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.820501] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1320.820501] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bfbca9-ab38-99f1-535e-e45d26023a71" [ 1320.820501] env[62952]: _type = "Task" [ 1320.820501] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.827389] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52bfbca9-ab38-99f1-535e-e45d26023a71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.254016] env[62952]: DEBUG nova.network.neutron [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updated VIF entry in instance network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1321.254439] env[62952]: DEBUG nova.network.neutron [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.331150] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52bfbca9-ab38-99f1-535e-e45d26023a71, 'name': SearchDatastore_Task, 'duration_secs': 0.00985} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.331463] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1321.331698] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Processing image e04fcbd0-b3d8-461a-890d-eda04a2a1e77 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1321.331933] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1321.332102] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.332289] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1321.332548] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-675bf8c7-d1e4-43e9-a529-8777b13feba9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.340216] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1321.340397] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1321.341097] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c266c84-c124-4881-a7ef-a9ab8429303f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.345797] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1321.345797] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52be0d2d-933e-f005-7bcb-5e3fc223b95e" [ 1321.345797] env[62952]: _type = "Task" [ 1321.345797] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.352900] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52be0d2d-933e-f005-7bcb-5e3fc223b95e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.757741] env[62952]: DEBUG oslo_concurrency.lockutils [req-b8b1dd7f-1caa-4ba0-800a-0cb15a009241 req-047da422-6fdd-4fed-ae24-9d1f4c7579db service nova] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1321.855956] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52be0d2d-933e-f005-7bcb-5e3fc223b95e, 'name': SearchDatastore_Task, 'duration_secs': 0.007835} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.856722] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72881c6c-1b86-4aa9-862a-56bbaad55896 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.861357] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1321.861357] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]5252df39-cfd0-dadd-f631-0dce6ef752cf" [ 1321.861357] env[62952]: _type = "Task" [ 1321.861357] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.869057] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5252df39-cfd0-dadd-f631-0dce6ef752cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.372052] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]5252df39-cfd0-dadd-f631-0dce6ef752cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009023} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.372052] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1322.372052] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1322.372319] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-436483c9-607a-4c2b-b4c3-d35980c9d265 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.379377] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1322.379377] env[62952]: value = "task-1367889" [ 1322.379377] env[62952]: _type = "Task" [ 1322.379377] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.387132] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.889426] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445506} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.889805] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e04fcbd0-b3d8-461a-890d-eda04a2a1e77/e04fcbd0-b3d8-461a-890d-eda04a2a1e77.vmdk to [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1322.889912] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Extending root virtual disk to 1048576 {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1322.890186] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c59c103-67b5-424a-a4e7-b8f54cc2cee7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.897127] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1322.897127] env[62952]: value = "task-1367890" [ 1322.897127] env[62952]: _type = "Task" [ 1322.897127] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.904108] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.407224] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055039} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.407500] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Extended root virtual disk {{(pid=62952) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1323.408246] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445def8f-bf5e-4948-b8bd-1a5b5729f1a3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.430336] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1323.430566] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72384b0b-ea15-4444-a2e6-9ec0a969fb48 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.449788] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1323.449788] env[62952]: value = "task-1367891" [ 1323.449788] env[62952]: _type = "Task" [ 1323.449788] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.457328] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367891, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.965074] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367891, 'name': ReconfigVM_Task, 'duration_secs': 0.278932} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.965489] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk or device None with type sparse {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1323.966183] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b334015-b7b0-480f-b315-1e82dbff09c6 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.973328] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1323.973328] env[62952]: value = "task-1367892" [ 1323.973328] env[62952]: _type = "Task" [ 1323.973328] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.981406] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367892, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.483246] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367892, 'name': Rename_Task, 'duration_secs': 0.15076} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.483542] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1324.483790] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-850280b2-473f-46ae-b8ee-ab224c6c8ae0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.490527] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1324.490527] env[62952]: value = "task-1367893" [ 1324.490527] env[62952]: _type = "Task" [ 1324.490527] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.498042] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.000422] env[62952]: DEBUG oslo_vmware.api [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367893, 'name': PowerOnVM_Task, 'duration_secs': 0.438169} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.000805] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1325.000929] env[62952]: INFO nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Took 6.61 seconds to spawn the instance on the hypervisor. [ 1325.001060] env[62952]: DEBUG nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1325.001799] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec64311-f7a1-412e-b536-74da9a68d93d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.517915] env[62952]: INFO nova.compute.manager [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Took 11.26 seconds to build instance. [ 1325.715293] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.019936] env[62952]: DEBUG oslo_concurrency.lockutils [None req-16f5996e-66ff-4b0d-8ad4-5bc63d127f05 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.773s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.218094] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.218334] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.221849] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.221849] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1326.221849] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d07c3b-1fab-4deb-8ef9-c167601bfada {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.228272] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213d6c53-8013-466d-a9d9-8052ae54db15 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.244083] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896a1e84-b92a-4dfd-a466-9cf7bd6fd55a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.247255] env[62952]: DEBUG nova.compute.manager [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1326.247447] env[62952]: DEBUG nova.compute.manager [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing instance network info cache due to event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1326.247666] env[62952]: DEBUG oslo_concurrency.lockutils [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1326.247811] env[62952]: DEBUG oslo_concurrency.lockutils [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.247979] env[62952]: DEBUG nova.network.neutron [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1326.255522] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2554ab2-d36a-4826-86fe-ab2250bb5ba9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.286065] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180909MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1326.286279] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.286429] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.949514] env[62952]: DEBUG nova.network.neutron [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updated VIF entry in instance network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1326.949841] env[62952]: DEBUG nova.network.neutron [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1327.311944] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 2352cb5e-f9e9-49a7-b595-617ef1bacda1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1327.312193] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1327.312319] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1327.338452] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f0d13a-23c6-4c71-abc6-5a445ea145b5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.345808] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ee4ef3-4f4e-429e-86d3-cc44d4474c75 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.376301] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b7961c-9e5b-4ee9-b45d-db5c3b175c40 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.383710] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6deca059-d82f-4425-b02c-b39334cac87a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.396466] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1327.452952] env[62952]: DEBUG oslo_concurrency.lockutils [req-071f3e29-de14-4ab3-9c8c-f1df488dfe68 req-8b1b5fa1-216f-42a3-98a7-86c98e3b89c6 service nova] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1327.900710] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1328.407176] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1328.407488] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.121s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1329.407488] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.407886] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.407886] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.408050] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.408216] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.408462] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1330.715019] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.715422] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1330.715422] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1331.247570] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1331.247715] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1331.247873] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1331.248043] env[62952]: DEBUG nova.objects.instance [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lazy-loading 'info_cache' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1332.952350] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1333.454854] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1333.495319] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1333.495319] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1333.495319] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1364.103249] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1364.103596] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1364.103725] env[62952]: INFO nova.compute.manager [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Shelving [ 1364.612065] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1364.612365] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db13b91e-2219-467c-868a-fb97ea9dfeae {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.620624] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1364.620624] env[62952]: value = "task-1367894" [ 1364.620624] env[62952]: _type = "Task" [ 1364.620624] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.628950] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367894, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.130606] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367894, 'name': PowerOffVM_Task, 'duration_secs': 0.162035} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.130957] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1365.131638] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c490557-c211-416e-a632-a2f4ff22d2e0 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.149722] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6a296d-e716-4c4f-8494-ef13f17cf6c8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.659630] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Creating Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1365.659946] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e2cbda66-9340-487b-9460-d7620638ae33 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.667415] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1365.667415] env[62952]: value = "task-1367895" [ 1365.667415] env[62952]: _type = "Task" [ 1365.667415] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.675397] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367895, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.177612] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367895, 'name': CreateSnapshot_Task, 'duration_secs': 0.386305} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.177990] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Created Snapshot of the VM instance {{(pid=62952) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1366.178702] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f167c66c-3462-47ca-b33a-2c17632a6acb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.696593] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Creating linked-clone VM from snapshot {{(pid=62952) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1366.696900] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-27149eb7-7e3a-4d68-9e86-8b8ed4182bc9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.705473] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1366.705473] env[62952]: value = "task-1367896" [ 1366.705473] env[62952]: _type = "Task" [ 1366.705473] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.713139] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367896, 'name': CloneVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.215983] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367896, 'name': CloneVM_Task} progress is 94%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.717495] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367896, 'name': CloneVM_Task, 'duration_secs': 0.880956} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.717750] env[62952]: INFO nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Created linked-clone VM from snapshot [ 1367.718487] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7542e45d-56ab-475f-b3f1-e57cbc471966 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.725177] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Uploading image de678e06-ce39-49e9-a69d-8a6b386eec09 {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1367.745190] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1367.745190] env[62952]: value = "vm-291013" [ 1367.745190] env[62952]: _type = "VirtualMachine" [ 1367.745190] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1367.745450] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-16217edf-8d62-4362-9eac-c06f9ce64050 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.751429] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease: (returnval){ [ 1367.751429] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bd56fd-6e02-9f21-e440-9bd33f9fc163" [ 1367.751429] env[62952]: _type = "HttpNfcLease" [ 1367.751429] env[62952]: } obtained for exporting VM: (result){ [ 1367.751429] env[62952]: value = "vm-291013" [ 1367.751429] env[62952]: _type = "VirtualMachine" [ 1367.751429] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1367.751661] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the lease: (returnval){ [ 1367.751661] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bd56fd-6e02-9f21-e440-9bd33f9fc163" [ 1367.751661] env[62952]: _type = "HttpNfcLease" [ 1367.751661] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1367.757527] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1367.757527] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bd56fd-6e02-9f21-e440-9bd33f9fc163" [ 1367.757527] env[62952]: _type = "HttpNfcLease" [ 1367.757527] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1368.259679] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1368.259679] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bd56fd-6e02-9f21-e440-9bd33f9fc163" [ 1368.259679] env[62952]: _type = "HttpNfcLease" [ 1368.259679] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1368.260193] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1368.260193] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52bd56fd-6e02-9f21-e440-9bd33f9fc163" [ 1368.260193] env[62952]: _type = "HttpNfcLease" [ 1368.260193] env[62952]: }. {{(pid=62952) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1368.260633] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c14049-d9b6-4542-86d1-07a266682934 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.267667] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5220f93d-8109-4408-e474-15a488714223/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1368.267842] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5220f93d-8109-4408-e474-15a488714223/disk-0.vmdk for reading. {{(pid=62952) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1368.352434] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a487e25d-f73a-4e88-af09-e323d09b5ca7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.518852] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5220f93d-8109-4408-e474-15a488714223/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1375.519793] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6c6e7a-5ee2-4ec1-bc1c-897a4c668fcd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.526036] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5220f93d-8109-4408-e474-15a488714223/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1375.526211] env[62952]: ERROR oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5220f93d-8109-4408-e474-15a488714223/disk-0.vmdk due to incomplete transfer. [ 1375.526421] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5dfa9e55-129b-4547-a16b-8dd1fd4baf05 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.534442] env[62952]: DEBUG oslo_vmware.rw_handles [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5220f93d-8109-4408-e474-15a488714223/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1375.534640] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Uploaded image de678e06-ce39-49e9-a69d-8a6b386eec09 to the Glance image server {{(pid=62952) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1375.536773] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Destroying the VM {{(pid=62952) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1375.536997] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-400a9ae3-8254-4e3f-9a38-2ebc3cec1ed9 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.542473] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1375.542473] env[62952]: value = "task-1367898" [ 1375.542473] env[62952]: _type = "Task" [ 1375.542473] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1375.549699] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367898, 'name': Destroy_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.052126] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367898, 'name': Destroy_Task, 'duration_secs': 0.338734} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.052388] env[62952]: INFO nova.virt.vmwareapi.vm_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Destroyed the VM [ 1376.052625] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deleting Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1376.052916] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ec5b2702-fcf4-4c1a-9579-fb0bccb36768 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.059307] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1376.059307] env[62952]: value = "task-1367899" [ 1376.059307] env[62952]: _type = "Task" [ 1376.059307] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.066265] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367899, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.568960] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367899, 'name': RemoveSnapshot_Task, 'duration_secs': 0.322732} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.569355] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deleted Snapshot of the VM instance {{(pid=62952) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1376.569509] env[62952]: DEBUG nova.compute.manager [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1376.570265] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0afd44-6316-4a31-8bae-509cd6214db8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.081996] env[62952]: INFO nova.compute.manager [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Shelve offloading [ 1377.083777] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1377.084038] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e601e5d7-f33f-4068-aeec-e810bc596c3c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.091562] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1377.091562] env[62952]: value = "task-1367900" [ 1377.091562] env[62952]: _type = "Task" [ 1377.091562] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.099390] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.601703] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] VM already powered off {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1377.602069] env[62952]: DEBUG nova.compute.manager [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1377.602608] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171c38d3-d422-4af4-b6ff-5a3fa9120120 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.607810] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.607976] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.608167] env[62952]: DEBUG nova.network.neutron [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1378.302965] env[62952]: DEBUG nova.network.neutron [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1378.805574] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1379.013891] env[62952]: DEBUG nova.compute.manager [req-8fa3bcc3-1789-477c-b07b-6ca5ee35f2f4 req-35806114-a0ed-4f67-9792-aa4b8590bea5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-vif-unplugged-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1379.014149] env[62952]: DEBUG oslo_concurrency.lockutils [req-8fa3bcc3-1789-477c-b07b-6ca5ee35f2f4 req-35806114-a0ed-4f67-9792-aa4b8590bea5 service nova] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1379.014353] env[62952]: DEBUG oslo_concurrency.lockutils [req-8fa3bcc3-1789-477c-b07b-6ca5ee35f2f4 req-35806114-a0ed-4f67-9792-aa4b8590bea5 service nova] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1379.014528] env[62952]: DEBUG oslo_concurrency.lockutils [req-8fa3bcc3-1789-477c-b07b-6ca5ee35f2f4 req-35806114-a0ed-4f67-9792-aa4b8590bea5 service nova] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1379.014699] env[62952]: DEBUG nova.compute.manager [req-8fa3bcc3-1789-477c-b07b-6ca5ee35f2f4 req-35806114-a0ed-4f67-9792-aa4b8590bea5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] No waiting events found dispatching network-vif-unplugged-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1379.014869] env[62952]: WARNING nova.compute.manager [req-8fa3bcc3-1789-477c-b07b-6ca5ee35f2f4 req-35806114-a0ed-4f67-9792-aa4b8590bea5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received unexpected event network-vif-unplugged-9b7beec0-076c-404c-864f-f225c46c9ac2 for instance with vm_state shelved and task_state shelving_offloading. [ 1379.110109] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1379.111008] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b8da8d-c6ae-4d66-a4c1-b7a77b59b33a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.118783] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1379.119059] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf416ae7-11de-4d07-88b5-eb3452016998 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.188944] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1379.189130] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1379.189316] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleting the datastore file [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1379.189574] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11eedfff-14ac-44cb-a359-53bfe6908ad2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.197127] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1379.197127] env[62952]: value = "task-1367902" [ 1379.197127] env[62952]: _type = "Task" [ 1379.197127] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.204249] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367902, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.707618] env[62952]: DEBUG oslo_vmware.api [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367902, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134367} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.707865] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1379.708067] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1379.708256] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1379.732628] env[62952]: INFO nova.scheduler.client.report [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted allocations for instance 2352cb5e-f9e9-49a7-b595-617ef1bacda1 [ 1380.237423] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1380.237789] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1380.237918] env[62952]: DEBUG nova.objects.instance [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'resources' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1380.740695] env[62952]: DEBUG nova.objects.instance [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'numa_topology' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1381.038566] env[62952]: DEBUG nova.compute.manager [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1381.038717] env[62952]: DEBUG nova.compute.manager [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing instance network info cache due to event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1381.038934] env[62952]: DEBUG oslo_concurrency.lockutils [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1381.039095] env[62952]: DEBUG oslo_concurrency.lockutils [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1381.039262] env[62952]: DEBUG nova.network.neutron [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1381.243444] env[62952]: DEBUG nova.objects.base [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Object Instance<2352cb5e-f9e9-49a7-b595-617ef1bacda1> lazy-loaded attributes: resources,numa_topology {{(pid=62952) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1381.258307] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf32a4cf-0deb-4637-8afd-5867eda5b538 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.265823] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a580b9-594d-4691-948f-88451008984e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.295397] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3773a7-d765-4658-b174-6bc412d01ece {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.303672] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0509e0-bff2-46be-8bf3-24c0e2ec9169 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.316463] env[62952]: DEBUG nova.compute.provider_tree [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1381.783782] env[62952]: DEBUG nova.network.neutron [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updated VIF entry in instance network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1381.784307] env[62952]: DEBUG nova.network.neutron [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9b7beec0-07", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1381.819629] env[62952]: DEBUG nova.scheduler.client.report [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1382.288435] env[62952]: DEBUG oslo_concurrency.lockutils [req-327e1906-23ea-4d10-85c9-4730e57a7793 req-ef22a51f-fa93-445f-b289-4b28f863061b service nova] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1382.324526] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1382.832928] env[62952]: DEBUG oslo_concurrency.lockutils [None req-be1a9ed2-4a18-41be-b2cf-1e15d94c6b72 tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.729s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1382.880645] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1382.880902] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1382.881097] env[62952]: INFO nova.compute.manager [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Unshelving [ 1383.902946] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1383.903272] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1383.903500] env[62952]: DEBUG nova.objects.instance [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'pci_requests' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1384.408073] env[62952]: DEBUG nova.objects.instance [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'numa_topology' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1384.911253] env[62952]: INFO nova.compute.claims [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1385.714220] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1385.946321] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6437dc31-ec8a-4c98-8edb-a96e18503029 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.953690] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfeb0ed-da28-459a-8146-08c0333b577e {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.982744] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ad946c-91e8-4e30-a1f3-da1e5fda1790 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.989166] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04118920-1c9f-4561-9c82-39f963c64956 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.001291] env[62952]: DEBUG nova.compute.provider_tree [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1386.217649] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1386.504099] env[62952]: DEBUG nova.scheduler.client.report [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1387.009456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.106s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1387.011638] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.794s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1387.011822] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1387.011985] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1387.013260] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395c3022-4248-4dd7-86b2-802f42a6a766 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.021472] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fae226-f5c1-4914-bbd0-a44967c77864 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.035982] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45469eec-4f69-4954-b7bf-9c15e424b56f {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.042025] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57af7532-92fe-4aed-80a6-146cb355aa22 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.045552] env[62952]: INFO nova.network.neutron [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating port 9b7beec0-076c-404c-864f-f225c46c9ac2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1387.072313] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181503MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1387.072456] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1387.072647] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1388.094394] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Instance 2352cb5e-f9e9-49a7-b595-617ef1bacda1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62952) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1388.094724] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1388.094810] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1388.119861] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4715be-a9bd-4483-a278-68f3335360c7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.127338] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246d34cc-e9ca-4798-87bd-185076c3ff29 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.156754] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c569ea69-289f-402d-bb2e-98700bf2d433 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.163237] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80924f07-2504-46e0-b066-776e1a7872e2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.175447] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1388.420032] env[62952]: DEBUG nova.compute.manager [req-e8c6899c-426b-4a8d-8b6e-98f2dcf3fc4a req-20d7bdc0-e501-48d4-8513-33ef39462627 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-vif-plugged-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1388.420737] env[62952]: DEBUG oslo_concurrency.lockutils [req-e8c6899c-426b-4a8d-8b6e-98f2dcf3fc4a req-20d7bdc0-e501-48d4-8513-33ef39462627 service nova] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1388.420966] env[62952]: DEBUG oslo_concurrency.lockutils [req-e8c6899c-426b-4a8d-8b6e-98f2dcf3fc4a req-20d7bdc0-e501-48d4-8513-33ef39462627 service nova] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1388.421194] env[62952]: DEBUG oslo_concurrency.lockutils [req-e8c6899c-426b-4a8d-8b6e-98f2dcf3fc4a req-20d7bdc0-e501-48d4-8513-33ef39462627 service nova] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1388.421418] env[62952]: DEBUG nova.compute.manager [req-e8c6899c-426b-4a8d-8b6e-98f2dcf3fc4a req-20d7bdc0-e501-48d4-8513-33ef39462627 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] No waiting events found dispatching network-vif-plugged-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1388.421595] env[62952]: WARNING nova.compute.manager [req-e8c6899c-426b-4a8d-8b6e-98f2dcf3fc4a req-20d7bdc0-e501-48d4-8513-33ef39462627 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received unexpected event network-vif-plugged-9b7beec0-076c-404c-864f-f225c46c9ac2 for instance with vm_state shelved_offloaded and task_state spawning. [ 1388.508138] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1388.508355] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1388.508527] env[62952]: DEBUG nova.network.neutron [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Building network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1388.678323] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1389.183344] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1389.183691] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.111s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1389.198300] env[62952]: DEBUG nova.network.neutron [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.701472] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1389.727470] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-26T01:32:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7a24e96031a2b7ff02f72515b04cc25c',container_format='bare',created_at=2024-10-26T01:46:50Z,direct_url=,disk_format='vmdk',id=de678e06-ce39-49e9-a69d-8a6b386eec09,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-672592619-shelved',owner='a2c2a97031df4b75b835a96b41793c74',properties=ImageMetaProps,protected=,size=31670272,status='active',tags=,updated_at=2024-10-26T01:47:02Z,virtual_size=,visibility=), allow threads: False {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1389.727737] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1389.727899] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image limits 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1389.728098] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Flavor pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1389.728256] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Image pref 0:0:0 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1389.728410] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62952) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1389.728617] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1389.728780] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1389.728949] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Got 1 possible topologies {{(pid=62952) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1389.729135] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1389.729316] env[62952]: DEBUG nova.virt.hardware [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62952) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1389.730177] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f3cc4d-b55d-4b6d-a0bd-d045fba5a571 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.738166] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c12c257-3362-4ebb-82db-ba3704bcc5d5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.751059] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:0a:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd22cb4ec-277f-41ee-8aba-b3d54442b93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b7beec0-076c-404c-864f-f225c46c9ac2', 'vif_model': 'vmxnet3'}] {{(pid=62952) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1389.758419] env[62952]: DEBUG oslo.service.loopingcall [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1389.758645] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Creating VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1389.758849] env[62952]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8fb9665-90f0-40ef-9b51-0838674bd190 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.776790] env[62952]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1389.776790] env[62952]: value = "task-1367903" [ 1389.776790] env[62952]: _type = "Task" [ 1389.776790] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.783715] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367903, 'name': CreateVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.288643] env[62952]: DEBUG oslo_vmware.api [-] Task: {'id': task-1367903, 'name': CreateVM_Task, 'duration_secs': 0.297604} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.289051] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Created VM on the ESX host {{(pid=62952) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1390.289413] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.289607] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.290026] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1390.290268] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c3b1e68-c09c-4d34-8e82-60524798b96d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.294693] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1390.294693] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52b30838-d37a-b8af-b87d-62d1191388b4" [ 1390.294693] env[62952]: _type = "Task" [ 1390.294693] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.302078] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52b30838-d37a-b8af-b87d-62d1191388b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.446553] env[62952]: DEBUG nova.compute.manager [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1390.446782] env[62952]: DEBUG nova.compute.manager [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing instance network info cache due to event network-changed-9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1390.447022] env[62952]: DEBUG oslo_concurrency.lockutils [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.447170] env[62952]: DEBUG oslo_concurrency.lockutils [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.447345] env[62952]: DEBUG nova.network.neutron [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Refreshing network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1390.805241] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1390.805531] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Processing image de678e06-ce39-49e9-a69d-8a6b386eec09 {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1390.805773] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.805925] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquired lock "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.806129] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1390.806383] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8822ff45-cb71-4853-ac06-a1ff93664016 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.815047] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1390.815234] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62952) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1390.815910] env[62952]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67a1fcff-a2ad-4261-bd77-5101a99657b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.820640] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1390.820640] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]52fbea5d-1810-2f20-0260-2eb2067f4305" [ 1390.820640] env[62952]: _type = "Task" [ 1390.820640] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.827777] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': session[5269618d-3803-3add-a4ce-a26ac946b237]52fbea5d-1810-2f20-0260-2eb2067f4305, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.134281] env[62952]: DEBUG nova.network.neutron [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updated VIF entry in instance network info cache for port 9b7beec0-076c-404c-864f-f225c46c9ac2. {{(pid=62952) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1391.134729] env[62952]: DEBUG nova.network.neutron [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1391.330977] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Preparing fetch location {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1391.331353] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Fetch image to [datastore2] OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c/OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c.vmdk {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1391.331446] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Downloading stream optimized image de678e06-ce39-49e9-a69d-8a6b386eec09 to [datastore2] OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c/OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c.vmdk on the data store datastore2 as vApp {{(pid=62952) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1391.331624] env[62952]: DEBUG nova.virt.vmwareapi.images [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Downloading image file data de678e06-ce39-49e9-a69d-8a6b386eec09 to the ESX as VM named 'OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c' {{(pid=62952) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1391.393811] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1391.393811] env[62952]: value = "resgroup-9" [ 1391.393811] env[62952]: _type = "ResourcePool" [ 1391.393811] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1391.394095] env[62952]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d1a91031-4383-4962-b173-d2f8862aaebd {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.414325] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease: (returnval){ [ 1391.414325] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528bb434-bd55-6e7f-f5e1-88088bcae05f" [ 1391.414325] env[62952]: _type = "HttpNfcLease" [ 1391.414325] env[62952]: } obtained for vApp import into resource pool (val){ [ 1391.414325] env[62952]: value = "resgroup-9" [ 1391.414325] env[62952]: _type = "ResourcePool" [ 1391.414325] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1391.414644] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the lease: (returnval){ [ 1391.414644] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528bb434-bd55-6e7f-f5e1-88088bcae05f" [ 1391.414644] env[62952]: _type = "HttpNfcLease" [ 1391.414644] env[62952]: } to be ready. {{(pid=62952) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1391.424565] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1391.424565] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528bb434-bd55-6e7f-f5e1-88088bcae05f" [ 1391.424565] env[62952]: _type = "HttpNfcLease" [ 1391.424565] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1391.637299] env[62952]: DEBUG oslo_concurrency.lockutils [req-a856d7d7-b2b3-4435-9b37-7a49c1bdda9b req-4647d245-a3c8-4e10-9add-14859562bef5 service nova] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1391.922785] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1391.922785] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528bb434-bd55-6e7f-f5e1-88088bcae05f" [ 1391.922785] env[62952]: _type = "HttpNfcLease" [ 1391.922785] env[62952]: } is initializing. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1392.179177] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.179433] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.179602] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1392.179738] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1392.423392] env[62952]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1392.423392] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528bb434-bd55-6e7f-f5e1-88088bcae05f" [ 1392.423392] env[62952]: _type = "HttpNfcLease" [ 1392.423392] env[62952]: } is ready. {{(pid=62952) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1392.423809] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1392.423809] env[62952]: value = "session[5269618d-3803-3add-a4ce-a26ac946b237]528bb434-bd55-6e7f-f5e1-88088bcae05f" [ 1392.423809] env[62952]: _type = "HttpNfcLease" [ 1392.423809] env[62952]: }. {{(pid=62952) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1392.424413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c495426c-cf28-409a-a719-c359c71903b7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.431081] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5245e93c-47fc-2283-8314-8ce5781ebb18/disk-0.vmdk from lease info. {{(pid=62952) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1392.431228] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating HTTP connection to write to file with size = 31670272 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5245e93c-47fc-2283-8314-8ce5781ebb18/disk-0.vmdk. {{(pid=62952) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1392.493429] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3326f6c2-979f-4527-8356-376b7df50feb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.683389] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1392.683599] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquired lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1392.683790] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Forcefully refreshing network info cache for instance {{(pid=62952) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1392.684047] env[62952]: DEBUG nova.objects.instance [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lazy-loading 'info_cache' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1393.590362] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Completed reading data from the image iterator. {{(pid=62952) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1393.590822] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5245e93c-47fc-2283-8314-8ce5781ebb18/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1393.591633] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cc7817-8e04-4580-907c-e1c27ec901e1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.599156] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5245e93c-47fc-2283-8314-8ce5781ebb18/disk-0.vmdk is in state: ready. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1393.599359] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5245e93c-47fc-2283-8314-8ce5781ebb18/disk-0.vmdk. {{(pid=62952) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1393.599628] env[62952]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e832fa9b-c9a3-4132-8c94-af890f43338a {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.788532] env[62952]: DEBUG oslo_vmware.rw_handles [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5245e93c-47fc-2283-8314-8ce5781ebb18/disk-0.vmdk. {{(pid=62952) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1393.788734] env[62952]: INFO nova.virt.vmwareapi.images [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Downloaded image file data de678e06-ce39-49e9-a69d-8a6b386eec09 [ 1393.789564] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd0b861-6711-476c-8139-efb501f77ca1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.804401] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-019f9002-5243-49cb-8fb0-de27b502f5f8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.829867] env[62952]: INFO nova.virt.vmwareapi.images [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] The imported VM was unregistered [ 1393.832034] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Caching image {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1393.832286] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Creating directory with path [datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1393.832533] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b69aa32-fd3d-4972-ba54-efbc1df136cc {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.841946] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Created directory with path [datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09 {{(pid=62952) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1393.842143] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c/OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c.vmdk to [datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk. {{(pid=62952) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1393.842368] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5f7a3913-6624-483c-9483-b68f91be37e3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.848195] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1393.848195] env[62952]: value = "task-1367906" [ 1393.848195] env[62952]: _type = "Task" [ 1393.848195] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.855615] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367906, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.359925] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367906, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.432137] env[62952]: DEBUG nova.network.neutron [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [{"id": "9b7beec0-076c-404c-864f-f225c46c9ac2", "address": "fa:16:3e:cc:0a:6a", "network": {"id": "53737dd6-3305-4f78-b229-ea9c9761fdc4", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1964594011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2c2a97031df4b75b835a96b41793c74", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b7beec0-07", "ovs_interfaceid": "9b7beec0-076c-404c-864f-f225c46c9ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1394.859576] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367906, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.935539] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Releasing lock "refresh_cache-2352cb5e-f9e9-49a7-b595-617ef1bacda1" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1394.935707] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updated the network info_cache for instance {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1394.935864] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.936066] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.936235] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.936390] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.936575] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.936692] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.936826] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1395.360068] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367906, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1395.859852] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367906, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.361022] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367906, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.22355} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.361022] env[62952]: INFO nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c/OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c.vmdk to [datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk. [ 1396.361022] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Cleaning up location [datastore2] OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c {{(pid=62952) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1396.361022] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_06c2cf3b-4950-4ee4-9ac1-a0ed97c5837c {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1396.361022] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3bac965-12f4-4bd2-8312-d1c36e0f7952 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.368155] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1396.368155] env[62952]: value = "task-1367907" [ 1396.368155] env[62952]: _type = "Task" [ 1396.368155] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.375458] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.877891] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032903} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.878287] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1396.878338] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Releasing lock "[datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk" {{(pid=62952) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1396.878580] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk to [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1396.878831] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d237df1d-91ae-494a-9ecc-786d1fa87ddf {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.885970] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1396.885970] env[62952]: value = "task-1367908" [ 1396.885970] env[62952]: _type = "Task" [ 1396.885970] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.892939] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.396238] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.898166] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1398.398880] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1398.899896] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.401064] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.467321] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.904897] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367908, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.557795} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1399.905347] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/de678e06-ce39-49e9-a69d-8a6b386eec09/de678e06-ce39-49e9-a69d-8a6b386eec09.vmdk to [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk {{(pid=62952) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1399.906413] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb05626-571c-4aa4-bba4-227cee5c8ce1 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.929705] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1399.929943] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37d8c0fe-ecd0-4ab5-aea1-680591577d39 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.948748] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1399.948748] env[62952]: value = "task-1367909" [ 1399.948748] env[62952]: _type = "Task" [ 1399.948748] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.956070] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367909, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.459398] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367909, 'name': ReconfigVM_Task, 'duration_secs': 0.269191} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.459662] env[62952]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1/2352cb5e-f9e9-49a7-b595-617ef1bacda1.vmdk or device None with type streamOptimized {{(pid=62952) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1400.460299] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1dd18da2-e3b2-4bf0-bc22-e0ac7bee0a7c {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.466760] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1400.466760] env[62952]: value = "task-1367910" [ 1400.466760] env[62952]: _type = "Task" [ 1400.466760] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.474164] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367910, 'name': Rename_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.978947] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367910, 'name': Rename_Task, 'duration_secs': 0.128607} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.978947] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powering on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1400.978947] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1587c72d-44e8-441f-a619-43f217513ad7 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.986856] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1400.986856] env[62952]: value = "task-1367911" [ 1400.986856] env[62952]: _type = "Task" [ 1400.986856] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.994677] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.497082] env[62952]: DEBUG oslo_vmware.api [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367911, 'name': PowerOnVM_Task, 'duration_secs': 0.423242} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.497375] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powered on the VM {{(pid=62952) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1401.588644] env[62952]: DEBUG nova.compute.manager [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Checking state {{(pid=62952) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1401.589594] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7da83f3-ab1b-4aec-80d6-7ae9b7f76640 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.105760] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5b03a442-1d45-429b-8a53-8aebcb8cc2cf tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.225s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1437.854058] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1437.854445] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1437.854547] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1437.854829] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1437.855059] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1437.858386] env[62952]: INFO nova.compute.manager [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Terminating instance [ 1437.860181] env[62952]: DEBUG nova.compute.manager [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Start destroying the instance on the hypervisor. {{(pid=62952) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1437.860412] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Destroying instance {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1437.861279] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc772204-6b91-46ab-81a2-7e50e642cb6b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.868930] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powering off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1437.869171] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e082bb9-50a2-4443-8cc1-a6694fdd2d4b {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.875113] env[62952]: DEBUG oslo_vmware.api [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1437.875113] env[62952]: value = "task-1367912" [ 1437.875113] env[62952]: _type = "Task" [ 1437.875113] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1437.882380] env[62952]: DEBUG oslo_vmware.api [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.384736] env[62952]: DEBUG oslo_vmware.api [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367912, 'name': PowerOffVM_Task, 'duration_secs': 0.172272} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1438.384999] env[62952]: DEBUG nova.virt.vmwareapi.vm_util [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Powered off the VM {{(pid=62952) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1438.385189] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Unregistering the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1438.385460] env[62952]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0eb4caac-80ff-4f23-985d-fd8ac246fb69 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.448311] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Unregistered the VM {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1438.448534] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deleting contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1438.448742] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleting the datastore file [datastore2] 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1438.449019] env[62952]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3875a43f-64e4-4132-bae7-3e178be14761 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.455184] env[62952]: DEBUG oslo_vmware.api [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for the task: (returnval){ [ 1438.455184] env[62952]: value = "task-1367914" [ 1438.455184] env[62952]: _type = "Task" [ 1438.455184] env[62952]: } to complete. {{(pid=62952) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1438.462602] env[62952]: DEBUG oslo_vmware.api [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1438.964816] env[62952]: DEBUG oslo_vmware.api [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Task: {'id': task-1367914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128822} completed successfully. {{(pid=62952) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1438.965222] env[62952]: DEBUG nova.virt.vmwareapi.ds_util [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted the datastore file {{(pid=62952) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1438.965272] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deleted contents of the VM from datastore datastore2 {{(pid=62952) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1438.965479] env[62952]: DEBUG nova.virt.vmwareapi.vmops [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance destroyed {{(pid=62952) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1438.965673] env[62952]: INFO nova.compute.manager [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1438.965922] env[62952]: DEBUG oslo.service.loopingcall [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62952) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1438.966136] env[62952]: DEBUG nova.compute.manager [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Deallocating network for instance {{(pid=62952) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1438.966231] env[62952]: DEBUG nova.network.neutron [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] deallocate_for_instance() {{(pid=62952) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1439.379797] env[62952]: DEBUG nova.compute.manager [req-21beb4fe-269b-4b89-a59a-40c53d14036b req-096dfe40-357e-40d1-82f7-c7ac25cd79ba service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Received event network-vif-deleted-9b7beec0-076c-404c-864f-f225c46c9ac2 {{(pid=62952) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1439.380019] env[62952]: INFO nova.compute.manager [req-21beb4fe-269b-4b89-a59a-40c53d14036b req-096dfe40-357e-40d1-82f7-c7ac25cd79ba service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Neutron deleted interface 9b7beec0-076c-404c-864f-f225c46c9ac2; detaching it from the instance and deleting it from the info cache [ 1439.380256] env[62952]: DEBUG nova.network.neutron [req-21beb4fe-269b-4b89-a59a-40c53d14036b req-096dfe40-357e-40d1-82f7-c7ac25cd79ba service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1439.858442] env[62952]: DEBUG nova.network.neutron [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Updating instance_info_cache with network_info: [] {{(pid=62952) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1439.882075] env[62952]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08eb9b83-5d8a-409e-8d59-2490fa04bbb3 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.894374] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c42635-5f67-4d41-8d50-d5887c315ea8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.918200] env[62952]: DEBUG nova.compute.manager [req-21beb4fe-269b-4b89-a59a-40c53d14036b req-096dfe40-357e-40d1-82f7-c7ac25cd79ba service nova] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Detach interface failed, port_id=9b7beec0-076c-404c-864f-f225c46c9ac2, reason: Instance 2352cb5e-f9e9-49a7-b595-617ef1bacda1 could not be found. {{(pid=62952) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1440.361497] env[62952]: INFO nova.compute.manager [-] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Took 1.40 seconds to deallocate network for instance. [ 1440.867948] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1440.868239] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1440.868469] env[62952]: DEBUG nova.objects.instance [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lazy-loading 'resources' on Instance uuid 2352cb5e-f9e9-49a7-b595-617ef1bacda1 {{(pid=62952) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1441.404516] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4955486-4a25-4a59-a48d-350bc88215b2 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.411822] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16064e2-1613-4d5a-b4e0-17395d5c5b24 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.439855] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b6806f-ca5f-4a13-853f-d1df20d3cb16 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.446299] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7655354b-0377-4419-a1ee-c04a0374da5d {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.458591] env[62952]: DEBUG nova.compute.provider_tree [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1441.961431] env[62952]: DEBUG nova.scheduler.client.report [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1442.468332] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1442.491521] env[62952]: INFO nova.scheduler.client.report [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Deleted allocations for instance 2352cb5e-f9e9-49a7-b595-617ef1bacda1 [ 1442.999932] env[62952]: DEBUG oslo_concurrency.lockutils [None req-5731ff79-8d35-4fc5-9bef-a03da49a516d tempest-AttachVolumeShelveTestJSON-1347149213 tempest-AttachVolumeShelveTestJSON-1347149213-project-member] Lock "2352cb5e-f9e9-49a7-b595-617ef1bacda1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.146s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1447.715052] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.218895] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.219077] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1448.219252] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1448.219407] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62952) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1448.220360] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484534c5-7f9f-4318-89c3-b7804e7ea5c5 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.228358] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748a170f-ac21-4377-85c7-5276f8b1f420 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.242427] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55a5fff-931a-4424-b32e-9ffe8eab1855 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.248609] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f45962-6171-469a-8a20-7e3823c62ab8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.277382] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181430MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=62952) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1448.277548] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.277762] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1449.298840] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1449.299145] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62952) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1449.312038] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c371b0f-4e83-440a-961f-53c4a7d106eb {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.319189] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1392790-2744-4df7-9ca9-ec988d0fd119 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.347618] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84d1909-a382-40cd-8c31-9beab2e2f355 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.354179] env[62952]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959d6f83-8ccd-4ed3-8fd7-9f3ab1994aa8 {{(pid=62952) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.366362] env[62952]: DEBUG nova.compute.provider_tree [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed in ProviderTree for provider: 9b9257d4-3be7-42a5-ae8c-67cf64823681 {{(pid=62952) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1449.870016] env[62952]: DEBUG nova.scheduler.client.report [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Inventory has not changed for provider 9b9257d4-3be7-42a5-ae8c-67cf64823681 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62952) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1450.375476] env[62952]: DEBUG nova.compute.resource_tracker [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62952) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1450.375869] env[62952]: DEBUG oslo_concurrency.lockutils [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.098s {{(pid=62952) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1452.370908] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1452.371307] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1452.371362] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1452.371483] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1452.371642] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1452.371790] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62952) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1452.715630] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1453.715468] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1453.715823] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Starting heal instance info cache {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1453.715892] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Rebuilding the list of instances to heal {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1454.218926] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Didn't find any instances for network info cache update. {{(pid=62952) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1454.219201] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1454.219390] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1454.219523] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Cleaning up deleted instances {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1454.725273] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] There are 11 instances to clean {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1454.725625] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 2352cb5e-f9e9-49a7-b595-617ef1bacda1] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1455.228505] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: c1c3d578-37e0-4823-9acf-612a09f512f7] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1455.731555] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: dd7d45f3-31bb-43c3-9317-2142d29b6ef5] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1456.234944] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 27e78e78-3a7f-4ad4-b939-438cb2b90b55] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1456.738915] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 6606e84c-36cd-4519-beff-096a8f1bc09e] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1457.242106] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 26fdfac2-4ea1-4822-b38c-d8f7388436be] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1457.745621] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 3cc1f262-3e94-4a7e-997c-fbd24bad0aa5] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1458.248814] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: ff8301b2-c5cb-4e46-8f5c-98577dfd4835] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1458.752025] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 287250a7-3354-49b7-9194-d88ae51afdaf] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1459.255868] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: 532a2bcd-5d6a-4fa9-abc9-a048cc915fda] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1459.758986] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] [instance: aef28168-98a7-4f65-80e7-731633339abf] Instance has had 0 of 5 cleanup attempts {{(pid=62952) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1460.262687] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1460.262864] env[62952]: DEBUG nova.compute.manager [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Cleaning up deleted instances with incomplete migration {{(pid=62952) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1460.765927] env[62952]: DEBUG oslo_service.periodic_task [None req-4ceec8ff-258c-4993-9242-99014e68a7b7 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62952) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}